Mar 13 14:05:00 crc systemd[1]: Starting Kubernetes Kubelet... Mar 13 14:05:00 crc restorecon[4758]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 14:05:00 crc restorecon[4758]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 13 14:05:01 crc kubenswrapper[4907]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 14:05:01 crc kubenswrapper[4907]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 13 14:05:01 crc kubenswrapper[4907]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 14:05:01 crc kubenswrapper[4907]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 14:05:01 crc kubenswrapper[4907]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 13 14:05:01 crc kubenswrapper[4907]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.538149 4907 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543755 4907 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543793 4907 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543797 4907 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543802 4907 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543806 4907 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543810 4907 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543813 4907 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543818 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543823 4907 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543827 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543831 4907 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543835 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543839 4907 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543844 4907 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543848 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543854 4907 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543872 4907 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543876 4907 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543895 4907 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543898 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543902 4907 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543906 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543909 4907 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543913 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543916 4907 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543920 4907 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543923 4907 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543929 4907 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543932 4907 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543936 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543939 4907 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543943 4907 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543947 4907 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543952 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543956 4907 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543960 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543964 4907 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543967 4907 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543971 4907 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543975 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543979 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543983 4907 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543986 4907 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543991 4907 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543995 4907 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.543999 4907 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544003 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544006 4907 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544011 4907 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544016 4907 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544020 4907 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544024 4907 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544028 4907 feature_gate.go:330] unrecognized feature gate: Example Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544032 4907 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544036 4907 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544040 4907 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544044 4907 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544050 4907 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544055 4907 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544059 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544064 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544068 4907 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544073 4907 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544078 4907 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544083 4907 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544087 4907 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544091 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544095 4907 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544099 4907 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544103 4907 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.544110 4907 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545089 4907 flags.go:64] FLAG: --address="0.0.0.0" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545103 4907 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545112 4907 flags.go:64] FLAG: --anonymous-auth="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545118 4907 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545125 4907 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545130 4907 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545137 4907 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545144 4907 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545150 4907 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545155 4907 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545162 4907 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545168 4907 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545173 4907 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545178 4907 flags.go:64] FLAG: --cgroup-root="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545182 4907 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545187 4907 flags.go:64] FLAG: --client-ca-file="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545191 4907 flags.go:64] FLAG: --cloud-config="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545195 4907 flags.go:64] FLAG: --cloud-provider="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545199 4907 flags.go:64] FLAG: --cluster-dns="[]" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545206 4907 flags.go:64] FLAG: --cluster-domain="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545210 4907 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545215 4907 flags.go:64] FLAG: --config-dir="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545219 4907 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545224 4907 flags.go:64] FLAG: --container-log-max-files="5" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545236 4907 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545241 4907 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545246 4907 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545250 4907 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545256 4907 flags.go:64] FLAG: --contention-profiling="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545260 4907 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545265 4907 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545269 4907 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545274 4907 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545280 4907 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545284 4907 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545289 4907 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545295 4907 flags.go:64] FLAG: --enable-load-reader="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545299 4907 flags.go:64] FLAG: --enable-server="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545304 4907 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545311 4907 flags.go:64] FLAG: --event-burst="100" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545317 4907 flags.go:64] FLAG: --event-qps="50" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545321 4907 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545326 4907 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545330 4907 flags.go:64] FLAG: --eviction-hard="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545337 4907 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545341 4907 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545346 4907 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545350 4907 flags.go:64] FLAG: --eviction-soft="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545355 4907 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545359 4907 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545364 4907 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545368 4907 flags.go:64] FLAG: --experimental-mounter-path="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545373 4907 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545377 4907 flags.go:64] FLAG: --fail-swap-on="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545381 4907 flags.go:64] FLAG: --feature-gates="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545388 4907 flags.go:64] FLAG: --file-check-frequency="20s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545392 4907 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545397 4907 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545402 4907 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545407 4907 flags.go:64] FLAG: --healthz-port="10248" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545412 4907 flags.go:64] FLAG: --help="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545416 4907 flags.go:64] FLAG: --hostname-override="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545420 4907 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545425 4907 flags.go:64] FLAG: --http-check-frequency="20s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545430 4907 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545434 4907 flags.go:64] FLAG: --image-credential-provider-config="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545438 4907 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545442 4907 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545446 4907 flags.go:64] FLAG: --image-service-endpoint="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545450 4907 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545454 4907 flags.go:64] FLAG: --kube-api-burst="100" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545459 4907 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545464 4907 flags.go:64] FLAG: --kube-api-qps="50" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545469 4907 flags.go:64] FLAG: --kube-reserved="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545473 4907 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545477 4907 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545482 4907 flags.go:64] FLAG: --kubelet-cgroups="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545486 4907 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545491 4907 flags.go:64] FLAG: --lock-file="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545495 4907 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545499 4907 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545504 4907 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545513 4907 flags.go:64] FLAG: --log-json-split-stream="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545517 4907 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545522 4907 flags.go:64] FLAG: --log-text-split-stream="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545526 4907 flags.go:64] FLAG: --logging-format="text" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545530 4907 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545534 4907 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545538 4907 flags.go:64] FLAG: --manifest-url="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545543 4907 flags.go:64] FLAG: --manifest-url-header="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545549 4907 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545553 4907 flags.go:64] FLAG: --max-open-files="1000000" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545559 4907 flags.go:64] FLAG: --max-pods="110" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545564 4907 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545568 4907 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545573 4907 flags.go:64] FLAG: --memory-manager-policy="None" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545577 4907 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545581 4907 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545585 4907 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545589 4907 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545603 4907 flags.go:64] FLAG: --node-status-max-images="50" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545608 4907 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545612 4907 flags.go:64] FLAG: --oom-score-adj="-999" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545617 4907 flags.go:64] FLAG: --pod-cidr="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545621 4907 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545630 4907 flags.go:64] FLAG: --pod-manifest-path="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545635 4907 flags.go:64] FLAG: --pod-max-pids="-1" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545640 4907 flags.go:64] FLAG: --pods-per-core="0" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545645 4907 flags.go:64] FLAG: --port="10250" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545650 4907 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545654 4907 flags.go:64] FLAG: --provider-id="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545659 4907 flags.go:64] FLAG: --qos-reserved="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545663 4907 flags.go:64] FLAG: --read-only-port="10255" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545668 4907 flags.go:64] FLAG: --register-node="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545672 4907 flags.go:64] FLAG: --register-schedulable="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545677 4907 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545690 4907 flags.go:64] FLAG: --registry-burst="10" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545694 4907 flags.go:64] FLAG: --registry-qps="5" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545699 4907 flags.go:64] FLAG: --reserved-cpus="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545704 4907 flags.go:64] FLAG: --reserved-memory="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545711 4907 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545715 4907 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545720 4907 flags.go:64] FLAG: --rotate-certificates="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545724 4907 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545729 4907 flags.go:64] FLAG: --runonce="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545733 4907 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545738 4907 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545742 4907 flags.go:64] FLAG: --seccomp-default="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545747 4907 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545751 4907 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545756 4907 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545761 4907 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545766 4907 flags.go:64] FLAG: --storage-driver-password="root" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545771 4907 flags.go:64] FLAG: --storage-driver-secure="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545776 4907 flags.go:64] FLAG: --storage-driver-table="stats" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545781 4907 flags.go:64] FLAG: --storage-driver-user="root" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545786 4907 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545830 4907 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545835 4907 flags.go:64] FLAG: --system-cgroups="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545841 4907 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545849 4907 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545855 4907 flags.go:64] FLAG: --tls-cert-file="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545860 4907 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545867 4907 flags.go:64] FLAG: --tls-min-version="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545872 4907 flags.go:64] FLAG: --tls-private-key-file="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545891 4907 flags.go:64] FLAG: --topology-manager-policy="none" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545897 4907 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545902 4907 flags.go:64] FLAG: --topology-manager-scope="container" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545908 4907 flags.go:64] FLAG: --v="2" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545916 4907 flags.go:64] FLAG: --version="false" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545924 4907 flags.go:64] FLAG: --vmodule="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545932 4907 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.545938 4907 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546076 4907 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546083 4907 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546088 4907 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546093 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546097 4907 feature_gate.go:330] unrecognized feature gate: Example Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546101 4907 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546105 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546110 4907 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546114 4907 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546118 4907 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546121 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546125 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546130 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546135 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546139 4907 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546143 4907 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546147 4907 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546152 4907 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546156 4907 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546162 4907 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546166 4907 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546171 4907 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546176 4907 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546180 4907 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546185 4907 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546189 4907 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546193 4907 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546197 4907 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546201 4907 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546206 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546210 4907 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546214 4907 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546218 4907 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546223 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546227 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546231 4907 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546236 4907 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546249 4907 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546254 4907 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546259 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546263 4907 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546267 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546271 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546276 4907 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546280 4907 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546284 4907 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546288 4907 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546292 4907 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546296 4907 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546300 4907 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546305 4907 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546309 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546314 4907 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546318 4907 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546322 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546326 4907 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546330 4907 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546335 4907 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546339 4907 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546344 4907 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546348 4907 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546352 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546356 4907 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546360 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546367 4907 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546371 4907 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546375 4907 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546379 4907 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546382 4907 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546386 4907 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.546391 4907 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.546405 4907 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.561237 4907 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.561311 4907 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561459 4907 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561485 4907 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561494 4907 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561505 4907 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561514 4907 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561523 4907 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561533 4907 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561544 4907 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561556 4907 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561565 4907 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561573 4907 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561582 4907 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561590 4907 feature_gate.go:330] unrecognized feature gate: Example Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561598 4907 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561606 4907 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561614 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561622 4907 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561630 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561637 4907 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561646 4907 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561653 4907 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561661 4907 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561669 4907 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561677 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561685 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561693 4907 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561700 4907 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561708 4907 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561716 4907 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561724 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561731 4907 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561739 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561747 4907 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561755 4907 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561765 4907 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561773 4907 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561783 4907 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561796 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561838 4907 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561848 4907 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561857 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561865 4907 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561874 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561910 4907 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561918 4907 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561927 4907 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561936 4907 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561945 4907 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561953 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561960 4907 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561971 4907 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561981 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.561989 4907 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562000 4907 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562009 4907 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562019 4907 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562028 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562035 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562044 4907 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562052 4907 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562061 4907 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562069 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562078 4907 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562086 4907 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562093 4907 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562102 4907 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562110 4907 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562117 4907 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562125 4907 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562133 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562143 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.562161 4907 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562422 4907 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562440 4907 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562455 4907 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562467 4907 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562478 4907 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562488 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562498 4907 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562507 4907 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562516 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562528 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562540 4907 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562551 4907 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562561 4907 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562572 4907 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562584 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562595 4907 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562606 4907 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562615 4907 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562624 4907 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562633 4907 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562642 4907 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562652 4907 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562660 4907 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562670 4907 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562680 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562691 4907 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562700 4907 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562708 4907 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562716 4907 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562725 4907 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562735 4907 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562746 4907 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562757 4907 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562766 4907 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562775 4907 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562784 4907 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562792 4907 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562804 4907 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562815 4907 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562826 4907 feature_gate.go:330] unrecognized feature gate: Example Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562837 4907 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562846 4907 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562856 4907 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562868 4907 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562927 4907 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562940 4907 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562950 4907 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562960 4907 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562970 4907 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562979 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562988 4907 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.562998 4907 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563010 4907 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563021 4907 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563031 4907 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563041 4907 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563050 4907 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563063 4907 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563072 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563080 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563090 4907 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563099 4907 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563108 4907 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563118 4907 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563128 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563138 4907 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563148 4907 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563158 4907 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563167 4907 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563176 4907 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.563187 4907 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.563204 4907 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.564680 4907 server.go:940] "Client rotation is on, will bootstrap in background" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.570125 4907 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.574172 4907 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.574311 4907 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.577550 4907 server.go:997] "Starting client certificate rotation" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.577579 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.577777 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.606679 4907 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.610346 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.612920 4907 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.630999 4907 log.go:25] "Validated CRI v1 runtime API" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.663170 4907 log.go:25] "Validated CRI v1 image API" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.665288 4907 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.669759 4907 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-13-14-00-27-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.669786 4907 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.684160 4907 manager.go:217] Machine: {Timestamp:2026-03-13 14:05:01.681465454 +0000 UTC m=+0.581253163 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:05bfbfa3-f04f-45ba-8233-a355fc774c24 BootID:4b182bae-7f2d-4b54-934c-85a9a277d132 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:67:c2:b6 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:67:c2:b6 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b4:40:fc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5e:74:67 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d2:3a:d8 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:9d:df:4c Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:55:05:15 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:86:43:25:c0:f9:2c Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3a:ab:56:05:7d:c5 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.684367 4907 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.684519 4907 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.686212 4907 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.686571 4907 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.686647 4907 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.687035 4907 topology_manager.go:138] "Creating topology manager with none policy" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.687055 4907 container_manager_linux.go:303] "Creating device plugin manager" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.687720 4907 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.687775 4907 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.688786 4907 state_mem.go:36] "Initialized new in-memory state store" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.688962 4907 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.692908 4907 kubelet.go:418] "Attempting to sync node with API server" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.692959 4907 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.693014 4907 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.693042 4907 kubelet.go:324] "Adding apiserver pod source" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.693062 4907 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.698131 4907 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.699128 4907 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.700916 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.701117 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.701374 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.703351 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.704106 4907 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706724 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706793 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706814 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706833 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706864 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706922 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706946 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706977 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.706999 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.707020 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.707048 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.707067 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.708174 4907 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.709195 4907 server.go:1280] "Started kubelet" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.710423 4907 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.710637 4907 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.711245 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:01 crc systemd[1]: Started Kubernetes Kubelet. Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.712252 4907 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.714448 4907 server.go:460] "Adding debug handlers to kubelet server" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.715076 4907 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.715235 4907 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.718184 4907 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.718228 4907 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.718312 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.718401 4907 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.718939 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.717722 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.719734 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.719857 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.721330 4907 factory.go:153] Registering CRI-O factory Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.721400 4907 factory.go:221] Registration of the crio container factory successfully Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.721591 4907 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.721617 4907 factory.go:55] Registering systemd factory Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.721633 4907 factory.go:221] Registration of the systemd container factory successfully Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.721667 4907 factory.go:103] Registering Raw factory Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.721692 4907 manager.go:1196] Started watching for new ooms in manager Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.722708 4907 manager.go:319] Starting recovery of all containers Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739589 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739713 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739753 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739799 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739837 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739867 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739951 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.739981 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740032 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740064 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740095 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740132 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740161 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740217 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740251 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740290 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740320 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740363 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740396 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740439 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740482 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740517 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740567 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740599 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740630 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740670 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740715 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740758 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740805 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740832 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740860 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740937 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.740982 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741026 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741064 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741098 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741139 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741171 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741209 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741248 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741281 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741324 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741363 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741396 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741447 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741480 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741517 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741557 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741585 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741622 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741651 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741694 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741736 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741791 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741846 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741910 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.741984 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742020 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742056 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742084 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742122 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742150 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742180 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742215 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742243 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742287 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742315 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742343 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742379 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742466 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742545 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742585 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742631 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742693 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742738 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742772 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742801 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742829 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742905 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.742937 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.743168 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.743354 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.743379 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.744008 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.744800 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.744892 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.744967 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.745027 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.749081 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.749148 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.749212 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.749273 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.749334 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.749445 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.749510 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.750549 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.750629 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.750689 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.750759 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.750825 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.750904 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.750972 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751032 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751104 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751190 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751263 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751334 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751404 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751469 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751537 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751599 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751666 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.751727 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754155 4907 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754205 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754230 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754254 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754276 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754297 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754316 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754333 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754351 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754371 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754387 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754403 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754421 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754436 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754455 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754470 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754486 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754503 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754517 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754533 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754549 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754567 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754586 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754604 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754620 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754637 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754675 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754706 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754726 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754742 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754759 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754773 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754788 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754803 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754818 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754832 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754848 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754863 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754924 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754943 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.754958 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755015 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755033 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755049 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755065 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755081 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755096 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755114 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755131 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755164 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755181 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755198 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755241 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755258 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755273 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755288 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755310 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755351 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755373 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755394 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755414 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755435 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755458 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755480 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755501 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755522 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755543 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755560 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755576 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755593 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755608 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755623 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755638 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755657 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755674 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755690 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755709 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755724 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755741 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755755 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755771 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755791 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755806 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755821 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755839 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755857 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755872 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755922 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755940 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755957 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755974 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.755989 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.756006 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.756022 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.756039 4907 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.756053 4907 reconstruct.go:97] "Volume reconstruction finished" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.756061 4907 reconciler.go:26] "Reconciler: start to sync state" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.767483 4907 manager.go:324] Recovery completed Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.777173 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.778020 4907 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.779264 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.779339 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.779363 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.780699 4907 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.780759 4907 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.780787 4907 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.780806 4907 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.780826 4907 state_mem.go:36] "Initialized new in-memory state store" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.780796 4907 kubelet.go:2335] "Starting kubelet main sync loop" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.781235 4907 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 13 14:05:01 crc kubenswrapper[4907]: W0313 14:05:01.781723 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.781809 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.803313 4907 policy_none.go:49] "None policy: Start" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.804384 4907 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.804416 4907 state_mem.go:35] "Initializing new in-memory state store" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.819044 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.859425 4907 manager.go:334] "Starting Device Plugin manager" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.859482 4907 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.859501 4907 server.go:79] "Starting device plugin registration server" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.859951 4907 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.859966 4907 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.860191 4907 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.860270 4907 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.860278 4907 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.871199 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.882081 4907 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.882189 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.883986 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.884029 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.884040 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.884184 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.884690 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.884775 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.885000 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.885022 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.885034 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.885156 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.885338 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.885389 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.886446 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.886473 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.886486 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.886568 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.886622 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.886638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.886909 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.887019 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.887100 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.887929 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.887944 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.887976 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.887993 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.887956 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.888041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.888180 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.888249 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.888272 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.888283 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.888563 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.888612 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.889406 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.889442 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.889458 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.889914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.889949 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.889962 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.890329 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.890362 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.891475 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.891513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.891526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.920532 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959073 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959138 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959165 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959187 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959211 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959234 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959255 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959277 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959298 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959322 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959363 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959384 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.959428 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.960073 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.961638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.961770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.961860 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:01 crc kubenswrapper[4907]: I0313 14:05:01.961987 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:01 crc kubenswrapper[4907]: E0313 14:05:01.962563 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.060930 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061293 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061474 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061365 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061836 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061943 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.061872 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062120 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062264 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062341 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062370 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062484 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062517 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062567 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062591 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062633 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062641 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062582 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062676 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062707 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062732 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062738 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062773 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062840 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.062936 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.063274 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.162946 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.164804 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.164943 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.165007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.165086 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:02 crc kubenswrapper[4907]: E0313 14:05:02.165452 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.232502 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.250873 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.267241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.285163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.291709 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:02 crc kubenswrapper[4907]: W0313 14:05:02.292576 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ee552ec2e48a5c108fcf9c3bd6a2c17717d5c5a0ed4e51c6d2f70b1cb7e877e9 WatchSource:0}: Error finding container ee552ec2e48a5c108fcf9c3bd6a2c17717d5c5a0ed4e51c6d2f70b1cb7e877e9: Status 404 returned error can't find the container with id ee552ec2e48a5c108fcf9c3bd6a2c17717d5c5a0ed4e51c6d2f70b1cb7e877e9 Mar 13 14:05:02 crc kubenswrapper[4907]: W0313 14:05:02.300397 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-59911775da384157205a39f316bd66024be8d1209fb0408de68c7e1cf48f1be5 WatchSource:0}: Error finding container 59911775da384157205a39f316bd66024be8d1209fb0408de68c7e1cf48f1be5: Status 404 returned error can't find the container with id 59911775da384157205a39f316bd66024be8d1209fb0408de68c7e1cf48f1be5 Mar 13 14:05:02 crc kubenswrapper[4907]: W0313 14:05:02.306092 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ced94ec9123ec94a405f76fd7f6f3e675d71f793fcdf2b7be742454ed1e341a7 WatchSource:0}: Error finding container ced94ec9123ec94a405f76fd7f6f3e675d71f793fcdf2b7be742454ed1e341a7: Status 404 returned error can't find the container with id ced94ec9123ec94a405f76fd7f6f3e675d71f793fcdf2b7be742454ed1e341a7 Mar 13 14:05:02 crc kubenswrapper[4907]: W0313 14:05:02.313325 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4e27ef75e42f8eb3a98d932ea2eb8bda47adf69857bbc4ea3eb50180d79615b2 WatchSource:0}: Error finding container 4e27ef75e42f8eb3a98d932ea2eb8bda47adf69857bbc4ea3eb50180d79615b2: Status 404 returned error can't find the container with id 4e27ef75e42f8eb3a98d932ea2eb8bda47adf69857bbc4ea3eb50180d79615b2 Mar 13 14:05:02 crc kubenswrapper[4907]: W0313 14:05:02.321288 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-7239ba4b92154e6432afbba93274a199559421fbbfe5479eaba9614832192193 WatchSource:0}: Error finding container 7239ba4b92154e6432afbba93274a199559421fbbfe5479eaba9614832192193: Status 404 returned error can't find the container with id 7239ba4b92154e6432afbba93274a199559421fbbfe5479eaba9614832192193 Mar 13 14:05:02 crc kubenswrapper[4907]: E0313 14:05:02.321413 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.566573 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.568651 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.568700 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.568711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.568740 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:02 crc kubenswrapper[4907]: E0313 14:05:02.569358 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.712449 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.787773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ced94ec9123ec94a405f76fd7f6f3e675d71f793fcdf2b7be742454ed1e341a7"} Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.789424 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"59911775da384157205a39f316bd66024be8d1209fb0408de68c7e1cf48f1be5"} Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.792446 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ee552ec2e48a5c108fcf9c3bd6a2c17717d5c5a0ed4e51c6d2f70b1cb7e877e9"} Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.793862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7239ba4b92154e6432afbba93274a199559421fbbfe5479eaba9614832192193"} Mar 13 14:05:02 crc kubenswrapper[4907]: I0313 14:05:02.795079 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4e27ef75e42f8eb3a98d932ea2eb8bda47adf69857bbc4ea3eb50180d79615b2"} Mar 13 14:05:02 crc kubenswrapper[4907]: W0313 14:05:02.982613 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:02 crc kubenswrapper[4907]: E0313 14:05:02.982709 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:03 crc kubenswrapper[4907]: W0313 14:05:03.081956 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:03 crc kubenswrapper[4907]: E0313 14:05:03.082275 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:03 crc kubenswrapper[4907]: E0313 14:05:03.123025 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Mar 13 14:05:03 crc kubenswrapper[4907]: W0313 14:05:03.141751 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:03 crc kubenswrapper[4907]: E0313 14:05:03.141855 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:03 crc kubenswrapper[4907]: W0313 14:05:03.202712 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:03 crc kubenswrapper[4907]: E0313 14:05:03.202800 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.370179 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.371844 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.371902 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.371914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.371946 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:03 crc kubenswrapper[4907]: E0313 14:05:03.372457 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.712015 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.714283 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:05:03 crc kubenswrapper[4907]: E0313 14:05:03.715107 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.801749 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d" exitCode=0 Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.801829 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.802023 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.803646 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.803705 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.803728 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.804043 4907 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8" exitCode=0 Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.804163 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.804215 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.805450 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.805506 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.805534 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.806189 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.807612 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.807664 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.807683 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.807746 4907 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881" exitCode=0 Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.807830 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.807980 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.809511 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.809568 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.809589 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.812342 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.812400 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.812420 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.812430 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.812433 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.814020 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.814052 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.814069 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.815928 4907 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3" exitCode=0 Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.815988 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3"} Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.816081 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.818634 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.818677 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:03 crc kubenswrapper[4907]: I0313 14:05:03.818693 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.712825 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:04 crc kubenswrapper[4907]: E0313 14:05:04.724128 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Mar 13 14:05:04 crc kubenswrapper[4907]: W0313 14:05:04.797620 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:04 crc kubenswrapper[4907]: E0313 14:05:04.797727 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.830715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.830760 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.830785 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.830803 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.831722 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.831758 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.831771 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.836061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.836218 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.837434 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.837472 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.837485 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.842172 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.842209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.842228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.842243 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.846278 4907 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708" exitCode=0 Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.846355 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708"} Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.846523 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.847092 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.847752 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.847808 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.847824 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.851691 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.851841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.851858 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.973030 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.976470 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.976541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.976560 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:04 crc kubenswrapper[4907]: I0313 14:05:04.976598 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:04 crc kubenswrapper[4907]: E0313 14:05:04.977114 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.174:6443: connect: connection refused" node="crc" Mar 13 14:05:05 crc kubenswrapper[4907]: W0313 14:05:05.140616 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:05:05 crc kubenswrapper[4907]: E0313 14:05:05.140704 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.338940 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.785073 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.851650 4907 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb" exitCode=0 Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.851715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb"} Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.851842 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.853205 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.853259 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.853281 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.857407 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c"} Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.857472 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.857495 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.857511 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.857634 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.858994 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859050 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859074 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859136 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859160 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859170 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859264 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859306 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.859330 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.863195 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.863228 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:05 crc kubenswrapper[4907]: I0313 14:05:05.863239 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.763449 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.864998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb"} Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.865072 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.865074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209"} Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.865138 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.865193 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.865228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4"} Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.865261 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520"} Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.866384 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.866399 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.866432 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.866451 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.866477 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:06 crc kubenswrapper[4907]: I0313 14:05:06.866513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.794484 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.875636 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.875970 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa"} Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.876130 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.876772 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.876818 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.876834 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.877844 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.877924 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:07 crc kubenswrapper[4907]: I0313 14:05:07.877941 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.178142 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.179996 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.180034 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.180047 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.180074 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.878066 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.879097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.879129 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:08 crc kubenswrapper[4907]: I0313 14:05:08.879139 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.332967 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.333219 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.334924 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.334986 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.335011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.614849 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.615126 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.619684 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.619755 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:09 crc kubenswrapper[4907]: I0313 14:05:09.619784 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.074524 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.074778 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.076420 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.076612 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.076851 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.086335 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.533374 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.883552 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.885180 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.885260 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:10 crc kubenswrapper[4907]: I0313 14:05:10.885284 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.470059 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.470294 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.472026 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.472074 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.472084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:11 crc kubenswrapper[4907]: E0313 14:05:11.871413 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.886204 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.888011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.888088 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:11 crc kubenswrapper[4907]: I0313 14:05:11.888107 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:12 crc kubenswrapper[4907]: I0313 14:05:12.333534 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:12 crc kubenswrapper[4907]: I0313 14:05:12.334181 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:12 crc kubenswrapper[4907]: I0313 14:05:12.498305 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 13 14:05:12 crc kubenswrapper[4907]: I0313 14:05:12.499125 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:12 crc kubenswrapper[4907]: I0313 14:05:12.500635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:12 crc kubenswrapper[4907]: I0313 14:05:12.500742 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:12 crc kubenswrapper[4907]: I0313 14:05:12.500765 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.344616 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.344730 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.348568 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.348621 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.348648 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:15 crc kubenswrapper[4907]: W0313 14:05:15.482939 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.483049 4907 trace.go:236] Trace[1286910631]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Mar-2026 14:05:05.481) (total time: 10001ms): Mar 13 14:05:15 crc kubenswrapper[4907]: Trace[1286910631]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:05:15.482) Mar 13 14:05:15 crc kubenswrapper[4907]: Trace[1286910631]: [10.001499423s] [10.001499423s] END Mar 13 14:05:15 crc kubenswrapper[4907]: E0313 14:05:15.483075 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.573488 4907 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46124->192.168.126.11:17697: read: connection reset by peer" start-of-body= Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.573585 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46124->192.168.126.11:17697: read: connection reset by peer" Mar 13 14:05:15 crc kubenswrapper[4907]: W0313 14:05:15.657170 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.657345 4907 trace.go:236] Trace[431240111]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Mar-2026 14:05:05.655) (total time: 10001ms): Mar 13 14:05:15 crc kubenswrapper[4907]: Trace[431240111]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:05:15.657) Mar 13 14:05:15 crc kubenswrapper[4907]: Trace[431240111]: [10.001747778s] [10.001747778s] END Mar 13 14:05:15 crc kubenswrapper[4907]: E0313 14:05:15.657385 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.712619 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.824473 4907 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.824544 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.896557 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.899549 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c" exitCode=255 Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.899603 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c"} Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.899745 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.900566 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.900614 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.900632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:15 crc kubenswrapper[4907]: I0313 14:05:15.901293 4907 scope.go:117] "RemoveContainer" containerID="fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c" Mar 13 14:05:16 crc kubenswrapper[4907]: W0313 14:05:16.064732 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.064812 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.068329 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.070169 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.072793 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:16 crc kubenswrapper[4907]: W0313 14:05:16.075506 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.075575 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.079317 4907 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.079374 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.081474 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.085921 4907 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.085982 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.716480 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:16Z is after 2026-02-23T05:33:13Z Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.904581 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.905484 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.907864 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" exitCode=255 Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.907941 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33"} Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.908047 4907 scope.go:117] "RemoveContainer" containerID="fe6cd60e8033e3165c90914a90e482e2378e469fd98044ddc2488e020f75e14c" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.908156 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.909351 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.909407 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.909428 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:16 crc kubenswrapper[4907]: I0313 14:05:16.918995 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:16 crc kubenswrapper[4907]: E0313 14:05:16.919706 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:17 crc kubenswrapper[4907]: I0313 14:05:17.717764 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:17Z is after 2026-02-23T05:33:13Z Mar 13 14:05:17 crc kubenswrapper[4907]: I0313 14:05:17.911489 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 14:05:18 crc kubenswrapper[4907]: I0313 14:05:18.716252 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:18Z is after 2026-02-23T05:33:13Z Mar 13 14:05:19 crc kubenswrapper[4907]: W0313 14:05:19.571090 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:19Z is after 2026-02-23T05:33:13Z Mar 13 14:05:19 crc kubenswrapper[4907]: E0313 14:05:19.571211 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:19Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.623301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.623510 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.629160 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.629257 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.629285 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.630273 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:19 crc kubenswrapper[4907]: E0313 14:05:19.630654 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.630729 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.715482 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:19Z is after 2026-02-23T05:33:13Z Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.918473 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919295 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919334 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919346 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:19 crc kubenswrapper[4907]: I0313 14:05:19.919921 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:19 crc kubenswrapper[4907]: E0313 14:05:19.920083 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:20 crc kubenswrapper[4907]: I0313 14:05:20.717247 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:20Z is after 2026-02-23T05:33:13Z Mar 13 14:05:20 crc kubenswrapper[4907]: W0313 14:05:20.974674 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:20Z is after 2026-02-23T05:33:13Z Mar 13 14:05:20 crc kubenswrapper[4907]: E0313 14:05:20.974799 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.501035 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.501230 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.502567 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.502637 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.502656 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.520139 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.717311 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:21Z is after 2026-02-23T05:33:13Z Mar 13 14:05:21 crc kubenswrapper[4907]: E0313 14:05:21.871662 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.924306 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.925443 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.925507 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:21 crc kubenswrapper[4907]: I0313 14:05:21.925520 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.334152 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.334240 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.473667 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:22 crc kubenswrapper[4907]: E0313 14:05:22.474499 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:22Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475126 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475140 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.475166 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:22 crc kubenswrapper[4907]: E0313 14:05:22.478179 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:22Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:22 crc kubenswrapper[4907]: I0313 14:05:22.714989 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:22Z is after 2026-02-23T05:33:13Z Mar 13 14:05:23 crc kubenswrapper[4907]: I0313 14:05:23.716990 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:23Z is after 2026-02-23T05:33:13Z Mar 13 14:05:24 crc kubenswrapper[4907]: I0313 14:05:24.718011 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:24Z is after 2026-02-23T05:33:13Z Mar 13 14:05:24 crc kubenswrapper[4907]: I0313 14:05:24.855113 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:05:24 crc kubenswrapper[4907]: E0313 14:05:24.861316 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.164996 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.165315 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167190 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167244 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167258 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.167948 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:25 crc kubenswrapper[4907]: E0313 14:05:25.168158 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.717342 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:25Z is after 2026-02-23T05:33:13Z Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.824482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.934521 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936096 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:25 crc kubenswrapper[4907]: I0313 14:05:25.936825 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:25 crc kubenswrapper[4907]: E0313 14:05:25.937052 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:26 crc kubenswrapper[4907]: W0313 14:05:26.083159 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z Mar 13 14:05:26 crc kubenswrapper[4907]: E0313 14:05:26.083261 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:26 crc kubenswrapper[4907]: E0313 14:05:26.085706 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:26 crc kubenswrapper[4907]: I0313 14:05:26.717760 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z Mar 13 14:05:26 crc kubenswrapper[4907]: W0313 14:05:26.775387 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z Mar 13 14:05:26 crc kubenswrapper[4907]: E0313 14:05:26.775494 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:27 crc kubenswrapper[4907]: W0313 14:05:27.179577 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:27Z is after 2026-02-23T05:33:13Z Mar 13 14:05:27 crc kubenswrapper[4907]: E0313 14:05:27.179658 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:27Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:27 crc kubenswrapper[4907]: I0313 14:05:27.718161 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:27Z is after 2026-02-23T05:33:13Z Mar 13 14:05:28 crc kubenswrapper[4907]: W0313 14:05:28.384764 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:28Z is after 2026-02-23T05:33:13Z Mar 13 14:05:28 crc kubenswrapper[4907]: E0313 14:05:28.384931 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:28Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:28 crc kubenswrapper[4907]: I0313 14:05:28.715385 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:28Z is after 2026-02-23T05:33:13Z Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.479311 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:29 crc kubenswrapper[4907]: E0313 14:05:29.480642 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:29Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.480875 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.480966 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.480984 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.481024 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:29 crc kubenswrapper[4907]: E0313 14:05:29.484399 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:29Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:29 crc kubenswrapper[4907]: I0313 14:05:29.717671 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:29Z is after 2026-02-23T05:33:13Z Mar 13 14:05:30 crc kubenswrapper[4907]: I0313 14:05:30.717229 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:30Z is after 2026-02-23T05:33:13Z Mar 13 14:05:31 crc kubenswrapper[4907]: I0313 14:05:31.717138 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z Mar 13 14:05:31 crc kubenswrapper[4907]: E0313 14:05:31.871807 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.333731 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.333821 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.333939 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.334170 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336181 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336264 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.336938 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.337184 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23" gracePeriod=30 Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.716756 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:32Z is after 2026-02-23T05:33:13Z Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.955536 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956372 4907 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23" exitCode=255 Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956407 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23"} Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956448 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c"} Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.956548 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.957805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.957830 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:32 crc kubenswrapper[4907]: I0313 14:05:32.957841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:33 crc kubenswrapper[4907]: I0313 14:05:33.717062 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:33Z is after 2026-02-23T05:33:13Z Mar 13 14:05:34 crc kubenswrapper[4907]: I0313 14:05:34.715856 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:34Z is after 2026-02-23T05:33:13Z Mar 13 14:05:35 crc kubenswrapper[4907]: I0313 14:05:35.715491 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:35Z is after 2026-02-23T05:33:13Z Mar 13 14:05:36 crc kubenswrapper[4907]: E0313 14:05:36.089429 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.485456 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:36 crc kubenswrapper[4907]: E0313 14:05:36.486653 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.487360 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.487613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.487747 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.488055 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:36 crc kubenswrapper[4907]: E0313 14:05:36.494148 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:36 crc kubenswrapper[4907]: I0313 14:05:36.717377 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:36Z is after 2026-02-23T05:33:13Z Mar 13 14:05:37 crc kubenswrapper[4907]: I0313 14:05:37.715238 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:37Z is after 2026-02-23T05:33:13Z Mar 13 14:05:38 crc kubenswrapper[4907]: I0313 14:05:38.717197 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:38Z is after 2026-02-23T05:33:13Z Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.333847 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.334113 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.335763 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.335848 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.335869 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:39 crc kubenswrapper[4907]: I0313 14:05:39.715245 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:39Z is after 2026-02-23T05:33:13Z Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.533492 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.533776 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.535528 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.535599 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.535617 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.717608 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:40Z is after 2026-02-23T05:33:13Z Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.782077 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.784267 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.784364 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.784385 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:40 crc kubenswrapper[4907]: I0313 14:05:40.785293 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.275945 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.282523 4907 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:41Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.283954 4907 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.716545 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:41Z is after 2026-02-23T05:33:13Z Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.871986 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.985495 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.986823 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.988787 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" exitCode=255 Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.988842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db"} Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.988917 4907 scope.go:117] "RemoveContainer" containerID="8cc2d976ae1bb0e2999831b35473eb89231620ed94a8938258a0fbbd5a909a33" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.989135 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.990708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.990749 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.990767 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:41 crc kubenswrapper[4907]: I0313 14:05:41.991558 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:05:41 crc kubenswrapper[4907]: E0313 14:05:41.991835 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.333941 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.334051 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.719009 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:42Z is after 2026-02-23T05:33:13Z Mar 13 14:05:42 crc kubenswrapper[4907]: I0313 14:05:42.995198 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:05:43 crc kubenswrapper[4907]: E0313 14:05:43.493261 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:43Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.494288 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.495846 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.496123 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.496305 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.496481 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:43 crc kubenswrapper[4907]: E0313 14:05:43.502468 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:43Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 14:05:43 crc kubenswrapper[4907]: I0313 14:05:43.718588 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:43Z is after 2026-02-23T05:33:13Z Mar 13 14:05:44 crc kubenswrapper[4907]: I0313 14:05:44.717476 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:44Z is after 2026-02-23T05:33:13Z Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.164807 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.165075 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.166802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.166851 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.166867 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.167798 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:05:45 crc kubenswrapper[4907]: E0313 14:05:45.168222 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.717731 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:45Z is after 2026-02-23T05:33:13Z Mar 13 14:05:45 crc kubenswrapper[4907]: W0313 14:05:45.741811 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:45Z is after 2026-02-23T05:33:13Z Mar 13 14:05:45 crc kubenswrapper[4907]: E0313 14:05:45.742208 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:45Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:45 crc kubenswrapper[4907]: I0313 14:05:45.824161 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.009176 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.010444 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.010510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.010527 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.011328 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:05:46 crc kubenswrapper[4907]: E0313 14:05:46.011618 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:05:46 crc kubenswrapper[4907]: E0313 14:05:46.094122 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:46 crc kubenswrapper[4907]: W0313 14:05:46.637210 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z Mar 13 14:05:46 crc kubenswrapper[4907]: E0313 14:05:46.637645 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:46 crc kubenswrapper[4907]: I0313 14:05:46.717081 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:46Z is after 2026-02-23T05:33:13Z Mar 13 14:05:47 crc kubenswrapper[4907]: I0313 14:05:47.715099 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:47Z is after 2026-02-23T05:33:13Z Mar 13 14:05:47 crc kubenswrapper[4907]: W0313 14:05:47.755471 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:47Z is after 2026-02-23T05:33:13Z Mar 13 14:05:47 crc kubenswrapper[4907]: E0313 14:05:47.755618 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 14:05:48 crc kubenswrapper[4907]: I0313 14:05:48.715291 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:48Z is after 2026-02-23T05:33:13Z Mar 13 14:05:49 crc kubenswrapper[4907]: I0313 14:05:49.719397 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:50 crc kubenswrapper[4907]: E0313 14:05:50.501185 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.503160 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505057 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505072 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.505101 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:50 crc kubenswrapper[4907]: E0313 14:05:50.510249 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:05:50 crc kubenswrapper[4907]: I0313 14:05:50.718379 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:50 crc kubenswrapper[4907]: W0313 14:05:50.762574 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 13 14:05:50 crc kubenswrapper[4907]: E0313 14:05:50.763752 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 14:05:51 crc kubenswrapper[4907]: I0313 14:05:51.718749 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:51 crc kubenswrapper[4907]: E0313 14:05:51.872157 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:05:52 crc kubenswrapper[4907]: I0313 14:05:52.333931 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:05:52 crc kubenswrapper[4907]: I0313 14:05:52.334044 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:05:52 crc kubenswrapper[4907]: I0313 14:05:52.717612 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:53 crc kubenswrapper[4907]: I0313 14:05:53.720533 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:54 crc kubenswrapper[4907]: I0313 14:05:54.718708 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.718515 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.790582 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.790772 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.792312 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.792364 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:55 crc kubenswrapper[4907]: I0313 14:05:55.792381 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.099110 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3243feae5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,LastTimestamp:2026-03-13 14:05:01.709142757 +0000 UTC m=+0.608930476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.106400 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.111643 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.117962 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.123371 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba32d9aa280 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.866082944 +0000 UTC m=+0.765870643,LastTimestamp:2026-03-13 14:05:01.866082944 +0000 UTC m=+0.765870643,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.128972 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.884013463 +0000 UTC m=+0.783801162,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.136161 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.884035894 +0000 UTC m=+0.783823593,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.142209 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.884047345 +0000 UTC m=+0.783835044,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.147541 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.885015553 +0000 UTC m=+0.784803242,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.152604 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.885029624 +0000 UTC m=+0.784817313,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.158359 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.885040344 +0000 UTC m=+0.784828033,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.164743 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.886466179 +0000 UTC m=+0.786253878,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.170773 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.88648071 +0000 UTC m=+0.786268409,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.176120 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.886494331 +0000 UTC m=+0.786282040,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.181539 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.886610818 +0000 UTC m=+0.786398507,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.188043 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.886631789 +0000 UTC m=+0.786419478,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.193670 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.88664583 +0000 UTC m=+0.786433519,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.198791 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.887948038 +0000 UTC m=+0.787735727,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.205215 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.887966539 +0000 UTC m=+0.787754248,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.208592 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.88798662 +0000 UTC m=+0.787774319,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.212601 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.888002191 +0000 UTC m=+0.787789900,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.219290 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.888032943 +0000 UTC m=+0.787820642,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.225559 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f9543\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f9543 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779375427 +0000 UTC m=+0.679163136,LastTimestamp:2026-03-13 14:05:01.888069705 +0000 UTC m=+0.787857404,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.231340 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286e93ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286e93ab default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779309483 +0000 UTC m=+0.679097192,LastTimestamp:2026-03-13 14:05:01.888267746 +0000 UTC m=+0.788055435,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.236067 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c6ba3286f4326\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c6ba3286f4326 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:01.779354406 +0000 UTC m=+0.679142115,LastTimestamp:2026-03-13 14:05:01.888279387 +0000 UTC m=+0.788067076,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.243524 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba347c65aa6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.30515575 +0000 UTC m=+1.204943489,LastTimestamp:2026-03-13 14:05:02.30515575 +0000 UTC m=+1.204943489,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.249702 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba347c69067 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.305169511 +0000 UTC m=+1.204957240,LastTimestamp:2026-03-13 14:05:02.305169511 +0000 UTC m=+1.204957240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.256445 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba34828b7f5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.311602165 +0000 UTC m=+1.211389844,LastTimestamp:2026-03-13 14:05:02.311602165 +0000 UTC m=+1.211389844,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.263537 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba348d5c81f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.322944031 +0000 UTC m=+1.222731730,LastTimestamp:2026-03-13 14:05:02.322944031 +0000 UTC m=+1.222731730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.270918 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba348f41381 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.324929409 +0000 UTC m=+1.224717118,LastTimestamp:2026-03-13 14:05:02.324929409 +0000 UTC m=+1.224717118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.278679 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba36f0ea4fc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.964204796 +0000 UTC m=+1.863992485,LastTimestamp:2026-03-13 14:05:02.964204796 +0000 UTC m=+1.863992485,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.285517 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba36f0f21ad openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.964236717 +0000 UTC m=+1.864024406,LastTimestamp:2026-03-13 14:05:02.964236717 +0000 UTC m=+1.864024406,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.292993 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba36f102c73 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.964305011 +0000 UTC m=+1.864092700,LastTimestamp:2026-03-13 14:05:02.964305011 +0000 UTC m=+1.864092700,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.300031 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba36f1bf973 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.965078387 +0000 UTC m=+1.864866076,LastTimestamp:2026-03-13 14:05:02.965078387 +0000 UTC m=+1.864866076,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.307506 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba36f264b72 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.965754738 +0000 UTC m=+1.865542417,LastTimestamp:2026-03-13 14:05:02.965754738 +0000 UTC m=+1.865542417,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.314417 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba36fca6fa4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.976511908 +0000 UTC m=+1.876299597,LastTimestamp:2026-03-13 14:05:02.976511908 +0000 UTC m=+1.876299597,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.321323 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba36fe206de openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.97805795 +0000 UTC m=+1.877845639,LastTimestamp:2026-03-13 14:05:02.97805795 +0000 UTC m=+1.877845639,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.328607 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba36fe78e27 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.978420263 +0000 UTC m=+1.878207952,LastTimestamp:2026-03-13 14:05:02.978420263 +0000 UTC m=+1.878207952,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.335148 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba3700096a8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.98006084 +0000 UTC m=+1.879848539,LastTimestamp:2026-03-13 14:05:02.98006084 +0000 UTC m=+1.879848539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.341598 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3700e41c6 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.980956614 +0000 UTC m=+1.880744303,LastTimestamp:2026-03-13 14:05:02.980956614 +0000 UTC m=+1.880744303,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.348281 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba370135a3a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.981290554 +0000 UTC m=+1.881078243,LastTimestamp:2026-03-13 14:05:02.981290554 +0000 UTC m=+1.881078243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.354637 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38202bf57 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.282192215 +0000 UTC m=+2.181979944,LastTimestamp:2026-03-13 14:05:03.282192215 +0000 UTC m=+2.181979944,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.360943 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba382c72e2a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.295065642 +0000 UTC m=+2.194853371,LastTimestamp:2026-03-13 14:05:03.295065642 +0000 UTC m=+2.194853371,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.367536 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba382e0cc04 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.296744452 +0000 UTC m=+2.196532171,LastTimestamp:2026-03-13 14:05:03.296744452 +0000 UTC m=+2.196532171,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.374604 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38f25706a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.502569578 +0000 UTC m=+2.402357287,LastTimestamp:2026-03-13 14:05:03.502569578 +0000 UTC m=+2.402357287,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.380797 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38fe96641 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.515412033 +0000 UTC m=+2.415199722,LastTimestamp:2026-03-13 14:05:03.515412033 +0000 UTC m=+2.415199722,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.387456 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38ff9afeb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.516479467 +0000 UTC m=+2.416267156,LastTimestamp:2026-03-13 14:05:03.516479467 +0000 UTC m=+2.416267156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.394107 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba39d27afab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.737597867 +0000 UTC m=+2.637385566,LastTimestamp:2026-03-13 14:05:03.737597867 +0000 UTC m=+2.637385566,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.400540 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba39dc723ec openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.748047852 +0000 UTC m=+2.647835551,LastTimestamp:2026-03-13 14:05:03.748047852 +0000 UTC m=+2.647835551,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.407073 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3a1379f21 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.805751073 +0000 UTC m=+2.705538772,LastTimestamp:2026-03-13 14:05:03.805751073 +0000 UTC m=+2.705538772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.412730 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3a14c8c73 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.807122547 +0000 UTC m=+2.706910266,LastTimestamp:2026-03-13 14:05:03.807122547 +0000 UTC m=+2.706910266,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.417563 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3a192ddc2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.811730882 +0000 UTC m=+2.711518601,LastTimestamp:2026-03-13 14:05:03.811730882 +0000 UTC m=+2.711518601,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.419792 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3a229341d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.821583389 +0000 UTC m=+2.721371078,LastTimestamp:2026-03-13 14:05:03.821583389 +0000 UTC m=+2.721371078,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.424759 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3afe0c887 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.051718279 +0000 UTC m=+2.951505968,LastTimestamp:2026-03-13 14:05:04.051718279 +0000 UTC m=+2.951505968,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.426152 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3aff60fc7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.053112775 +0000 UTC m=+2.952900464,LastTimestamp:2026-03-13 14:05:04.053112775 +0000 UTC m=+2.952900464,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.432105 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3b003b64a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.05400737 +0000 UTC m=+2.953795059,LastTimestamp:2026-03-13 14:05:04.05400737 +0000 UTC m=+2.953795059,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.437527 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3b006033b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.054158139 +0000 UTC m=+2.953945828,LastTimestamp:2026-03-13 14:05:04.054158139 +0000 UTC m=+2.953945828,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.443377 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c6ba3b0b1c9dc openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.065415644 +0000 UTC m=+2.965203333,LastTimestamp:2026-03-13 14:05:04.065415644 +0000 UTC m=+2.965203333,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.449748 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3b0feca2c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.070461996 +0000 UTC m=+2.970249685,LastTimestamp:2026-03-13 14:05:04.070461996 +0000 UTC m=+2.970249685,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.455787 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3b11202b6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.071721654 +0000 UTC m=+2.971509343,LastTimestamp:2026-03-13 14:05:04.071721654 +0000 UTC m=+2.971509343,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.461537 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3b12f7f7e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.073654142 +0000 UTC m=+2.973441831,LastTimestamp:2026-03-13 14:05:04.073654142 +0000 UTC m=+2.973441831,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.467936 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3b140496a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.07475441 +0000 UTC m=+2.974542099,LastTimestamp:2026-03-13 14:05:04.07475441 +0000 UTC m=+2.974542099,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.472141 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3b1a728cb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.081496267 +0000 UTC m=+2.981283946,LastTimestamp:2026-03-13 14:05:04.081496267 +0000 UTC m=+2.981283946,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.477943 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3bc954357 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.264872791 +0000 UTC m=+3.164660480,LastTimestamp:2026-03-13 14:05:04.264872791 +0000 UTC m=+3.164660480,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.484335 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3bd1fb389 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.273945481 +0000 UTC m=+3.173733170,LastTimestamp:2026-03-13 14:05:04.273945481 +0000 UTC m=+3.173733170,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.490272 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3bd410ea8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.276131496 +0000 UTC m=+3.175919185,LastTimestamp:2026-03-13 14:05:04.276131496 +0000 UTC m=+3.175919185,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.496431 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3bd57864d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.277603917 +0000 UTC m=+3.177391606,LastTimestamp:2026-03-13 14:05:04.277603917 +0000 UTC m=+3.177391606,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.502928 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3bdfba96a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.28836081 +0000 UTC m=+3.188148489,LastTimestamp:2026-03-13 14:05:04.28836081 +0000 UTC m=+3.188148489,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.509413 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3be0c23cd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.289440717 +0000 UTC m=+3.189228426,LastTimestamp:2026-03-13 14:05:04.289440717 +0000 UTC m=+3.189228426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.516638 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3caba6eeb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.502189803 +0000 UTC m=+3.401977492,LastTimestamp:2026-03-13 14:05:04.502189803 +0000 UTC m=+3.401977492,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.522598 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3cad24630 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.50375224 +0000 UTC m=+3.403539929,LastTimestamp:2026-03-13 14:05:04.50375224 +0000 UTC m=+3.403539929,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.528784 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3cc433def openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.527932911 +0000 UTC m=+3.427720600,LastTimestamp:2026-03-13 14:05:04.527932911 +0000 UTC m=+3.427720600,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.534820 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3cc5aaac0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.529468096 +0000 UTC m=+3.429255785,LastTimestamp:2026-03-13 14:05:04.529468096 +0000 UTC m=+3.429255785,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.539017 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c6ba3cc780e43 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.531394115 +0000 UTC m=+3.431181804,LastTimestamp:2026-03-13 14:05:04.531394115 +0000 UTC m=+3.431181804,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.545413 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d86f79fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.732158462 +0000 UTC m=+3.631946151,LastTimestamp:2026-03-13 14:05:04.732158462 +0000 UTC m=+3.631946151,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.551143 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d972c8eb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.749152491 +0000 UTC m=+3.648940210,LastTimestamp:2026-03-13 14:05:04.749152491 +0000 UTC m=+3.648940210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.556947 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d995e9f9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.751454713 +0000 UTC m=+3.651242422,LastTimestamp:2026-03-13 14:05:04.751454713 +0000 UTC m=+3.651242422,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.563870 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3dfb3175c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.854030172 +0000 UTC m=+3.753817871,LastTimestamp:2026-03-13 14:05:04.854030172 +0000 UTC m=+3.753817871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.570000 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3e890b760 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.00277232 +0000 UTC m=+3.902560009,LastTimestamp:2026-03-13 14:05:05.00277232 +0000 UTC m=+3.902560009,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.575512 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3e95ca04a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.016135754 +0000 UTC m=+3.915923443,LastTimestamp:2026-03-13 14:05:05.016135754 +0000 UTC m=+3.915923443,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.581010 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3ece1107b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.075146875 +0000 UTC m=+3.974934574,LastTimestamp:2026-03-13 14:05:05.075146875 +0000 UTC m=+3.974934574,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.587164 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba3ed9af87b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.087330427 +0000 UTC m=+3.987118116,LastTimestamp:2026-03-13 14:05:05.087330427 +0000 UTC m=+3.987118116,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.591622 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba41b616858 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:05.855309912 +0000 UTC m=+4.755097601,LastTimestamp:2026-03-13 14:05:05.855309912 +0000 UTC m=+4.755097601,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.596831 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba4263cdc2f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.037464111 +0000 UTC m=+4.937251800,LastTimestamp:2026-03-13 14:05:06.037464111 +0000 UTC m=+4.937251800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.601685 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba426d74abc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.047584956 +0000 UTC m=+4.947372645,LastTimestamp:2026-03-13 14:05:06.047584956 +0000 UTC m=+4.947372645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.608160 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba426e6b436 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.048594998 +0000 UTC m=+4.948382687,LastTimestamp:2026-03-13 14:05:06.048594998 +0000 UTC m=+4.948382687,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.612647 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43179cc54 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.226007124 +0000 UTC m=+5.125794813,LastTimestamp:2026-03-13 14:05:06.226007124 +0000 UTC m=+5.125794813,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.619250 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba4324778f4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.239486196 +0000 UTC m=+5.139273885,LastTimestamp:2026-03-13 14:05:06.239486196 +0000 UTC m=+5.139273885,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.625955 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43254cda0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.24035984 +0000 UTC m=+5.140147539,LastTimestamp:2026-03-13 14:05:06.24035984 +0000 UTC m=+5.140147539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.637690 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43e729543 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.443638083 +0000 UTC m=+5.343425772,LastTimestamp:2026-03-13 14:05:06.443638083 +0000 UTC m=+5.343425772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.643824 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43f6c95ed openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.460022253 +0000 UTC m=+5.359809952,LastTimestamp:2026-03-13 14:05:06.460022253 +0000 UTC m=+5.359809952,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.649958 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba43f811e8a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.461367946 +0000 UTC m=+5.361155635,LastTimestamp:2026-03-13 14:05:06.461367946 +0000 UTC m=+5.361155635,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.656330 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba44f92e19b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.730967451 +0000 UTC m=+5.630755180,LastTimestamp:2026-03-13 14:05:06.730967451 +0000 UTC m=+5.630755180,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.661330 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45065b402 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.744783874 +0000 UTC m=+5.644571573,LastTimestamp:2026-03-13 14:05:06.744783874 +0000 UTC m=+5.644571573,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.668310 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45074c74e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.745771854 +0000 UTC m=+5.645559583,LastTimestamp:2026-03-13 14:05:06.745771854 +0000 UTC m=+5.645559583,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.675620 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45e840115 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.981650709 +0000 UTC m=+5.881438398,LastTimestamp:2026-03-13 14:05:06.981650709 +0000 UTC m=+5.881438398,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.680022 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c6ba45f41ed95 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:06.994097557 +0000 UTC m=+5.893885286,LastTimestamp:2026-03-13 14:05:06.994097557 +0000 UTC m=+5.893885286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.688323 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d8c577e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,LastTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.694803 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d903f8e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,LastTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.702965 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-apiserver-crc.189c6ba65ea21eea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:46124->192.168.126.11:17697: read: connection reset by peer Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.573559018 +0000 UTC m=+14.473346717,LastTimestamp:2026-03-13 14:05:15.573559018 +0000 UTC m=+14.473346717,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.709571 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba65ea3190c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:46124->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.573623052 +0000 UTC m=+14.473410741,LastTimestamp:2026-03-13 14:05:15.573623052 +0000 UTC m=+14.473410741,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: I0313 14:05:56.715823 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.716259 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-apiserver-crc.189c6ba66d978dd1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.824524753 +0000 UTC m=+14.724312442,LastTimestamp:2026-03-13 14:05:15.824524753 +0000 UTC m=+14.724312442,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.720327 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba66d983a0e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:15.824568846 +0000 UTC m=+14.724356535,LastTimestamp:2026-03-13 14:05:15.824568846 +0000 UTC m=+14.724356535,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.726532 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c6ba3d995e9f9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba3d995e9f9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:04.751454713 +0000 UTC m=+3.651242422,LastTimestamp:2026-03-13 14:05:15.902518846 +0000 UTC m=+14.802306525,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.730245 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-apiserver-crc.189c6ba67cc80036 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 13 14:05:56 crc kubenswrapper[4907]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 14:05:56 crc kubenswrapper[4907]: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:16.079358006 +0000 UTC m=+14.979145695,LastTimestamp:2026-03-13 14:05:16.079358006 +0000 UTC m=+14.979145695,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.736474 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c6ba67cc89e45 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:16.079398469 +0000 UTC m=+14.979186158,LastTimestamp:2026-03-13 14:05:16.079398469 +0000 UTC m=+14.979186158,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.743754 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba7f1999a2d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:22.334218797 +0000 UTC m=+21.234006526,LastTimestamp:2026-03-13 14:05:22.334218797 +0000 UTC m=+21.234006526,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.749990 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba7f19a748c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:22.3342747 +0000 UTC m=+21.234062429,LastTimestamp:2026-03-13 14:05:22.3342747 +0000 UTC m=+21.234062429,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.756182 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d8c577e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d8c577e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,LastTimestamp:2026-03-13 14:05:32.333799643 +0000 UTC m=+31.233587372,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.763136 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d903f8e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d903f8e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,LastTimestamp:2026-03-13 14:05:32.333862304 +0000 UTC m=+31.233650023,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.767665 4907 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6baa45d248e7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:32.337154279 +0000 UTC m=+31.236941998,LastTimestamp:2026-03-13 14:05:32.337154279 +0000 UTC m=+31.236941998,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.773527 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba370135a3a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba370135a3a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:02.981290554 +0000 UTC m=+1.881078243,LastTimestamp:2026-03-13 14:05:32.461056377 +0000 UTC m=+31.360844106,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.777761 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba38202bf57\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba38202bf57 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.282192215 +0000 UTC m=+2.181979944,LastTimestamp:2026-03-13 14:05:32.706689196 +0000 UTC m=+31.606476925,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.792702 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba382c72e2a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba382c72e2a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:03.295065642 +0000 UTC m=+2.194853371,LastTimestamp:2026-03-13 14:05:32.720223796 +0000 UTC m=+31.620011525,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.800635 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d8c577e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d8c577e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334129022 +0000 UTC m=+11.233916791,LastTimestamp:2026-03-13 14:05:42.33401524 +0000 UTC m=+41.233802959,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.805714 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba59d903f8e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c6ba59d903f8e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:12.334385038 +0000 UTC m=+11.234172767,LastTimestamp:2026-03-13 14:05:42.334087382 +0000 UTC m=+41.233875101,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:05:56 crc kubenswrapper[4907]: E0313 14:05:56.810407 4907 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c6ba7f1999a2d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 14:05:56 crc kubenswrapper[4907]: &Event{ObjectMeta:{kube-controller-manager-crc.189c6ba7f1999a2d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 14:05:56 crc kubenswrapper[4907]: body: Mar 13 14:05:56 crc kubenswrapper[4907]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:05:22.334218797 +0000 UTC m=+21.234006526,LastTimestamp:2026-03-13 14:05:52.334014601 +0000 UTC m=+51.233802320,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 14:05:56 crc kubenswrapper[4907]: > Mar 13 14:05:57 crc kubenswrapper[4907]: E0313 14:05:57.508461 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.511494 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513279 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513329 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513396 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.513476 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:05:57 crc kubenswrapper[4907]: E0313 14:05:57.519134 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:05:57 crc kubenswrapper[4907]: I0313 14:05:57.719700 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:58 crc kubenswrapper[4907]: I0313 14:05:58.717296 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.339825 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.340089 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.341394 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.341425 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.341433 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.351140 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:05:59 crc kubenswrapper[4907]: I0313 14:05:59.720208 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.048830 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.049920 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.049953 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.049963 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.718477 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.782384 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.783592 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.783635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.783648 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:00 crc kubenswrapper[4907]: I0313 14:06:00.784316 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:06:00 crc kubenswrapper[4907]: E0313 14:06:00.784528 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:01 crc kubenswrapper[4907]: I0313 14:06:01.716711 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:01 crc kubenswrapper[4907]: E0313 14:06:01.872401 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:02 crc kubenswrapper[4907]: I0313 14:06:02.716998 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:03 crc kubenswrapper[4907]: I0313 14:06:03.717634 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:04 crc kubenswrapper[4907]: E0313 14:06:04.515662 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.520067 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521659 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521714 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521732 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.521763 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:04 crc kubenswrapper[4907]: E0313 14:06:04.528424 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:06:04 crc kubenswrapper[4907]: I0313 14:06:04.718175 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:05 crc kubenswrapper[4907]: I0313 14:06:05.715954 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:06 crc kubenswrapper[4907]: I0313 14:06:06.715471 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:07 crc kubenswrapper[4907]: I0313 14:06:07.717082 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:08 crc kubenswrapper[4907]: I0313 14:06:08.717806 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:09 crc kubenswrapper[4907]: I0313 14:06:09.718513 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:10 crc kubenswrapper[4907]: I0313 14:06:10.719468 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:11 crc kubenswrapper[4907]: E0313 14:06:11.523942 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.529074 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531109 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531151 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531163 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.531196 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:11 crc kubenswrapper[4907]: E0313 14:06:11.538168 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:06:11 crc kubenswrapper[4907]: I0313 14:06:11.717069 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:11 crc kubenswrapper[4907]: E0313 14:06:11.873335 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.718498 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.781746 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783039 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783079 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:12 crc kubenswrapper[4907]: I0313 14:06:12.783802 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.091671 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.093921 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80"} Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.094094 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.095098 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.095138 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.095148 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.286092 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.300093 4907 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 14:06:13 crc kubenswrapper[4907]: I0313 14:06:13.715373 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.098876 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.099696 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.101730 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" exitCode=255 Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.101782 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80"} Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.101827 4907 scope.go:117] "RemoveContainer" containerID="da9239d5489f168f1e0695a6eb1d540c331de494e29ae89624cfbfed390520db" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.102034 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.102951 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.102992 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.103007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.103708 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:14 crc kubenswrapper[4907]: E0313 14:06:14.103944 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:14 crc kubenswrapper[4907]: I0313 14:06:14.718620 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.108138 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.164624 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.164964 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.166169 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.166219 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.166234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.167044 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:15 crc kubenswrapper[4907]: E0313 14:06:15.167245 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:15 crc kubenswrapper[4907]: W0313 14:06:15.302390 4907 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:15 crc kubenswrapper[4907]: E0313 14:06:15.302501 4907 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.716786 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:15 crc kubenswrapper[4907]: I0313 14:06:15.824392 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.114455 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.115353 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.115413 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.115431 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.116408 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:16 crc kubenswrapper[4907]: E0313 14:06:16.116691 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:16 crc kubenswrapper[4907]: I0313 14:06:16.716383 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:17 crc kubenswrapper[4907]: I0313 14:06:17.716098 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:18 crc kubenswrapper[4907]: E0313 14:06:18.532060 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.538671 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540201 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540292 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540355 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.540398 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:18 crc kubenswrapper[4907]: E0313 14:06:18.548688 4907 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 14:06:18 crc kubenswrapper[4907]: I0313 14:06:18.719389 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.714977 4907 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.831796 4907 csr.go:261] certificate signing request csr-db9vj is approved, waiting to be issued Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.840220 4907 csr.go:257] certificate signing request csr-db9vj is issued Mar 13 14:06:19 crc kubenswrapper[4907]: I0313 14:06:19.935138 4907 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 13 14:06:20 crc kubenswrapper[4907]: I0313 14:06:20.577576 4907 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 13 14:06:20 crc kubenswrapper[4907]: I0313 14:06:20.842022 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-24 22:13:07.011008384 +0000 UTC Mar 13 14:06:20 crc kubenswrapper[4907]: I0313 14:06:20.842099 4907 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6152h6m46.168914302s for next certificate rotation Mar 13 14:06:21 crc kubenswrapper[4907]: E0313 14:06:21.874204 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.549656 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552031 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552122 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552148 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.552344 4907 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.564478 4907 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.564857 4907 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.564927 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570517 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570570 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570585 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570609 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.570626 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.590615 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602231 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602319 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.602369 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.618805 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630098 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630206 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630225 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.630271 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.648941 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660313 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660365 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660382 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660406 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.660426 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:25Z","lastTransitionTime":"2026-03-13T14:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.677815 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.678116 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.678166 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.779115 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.781498 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.783174 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.783227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:25 crc kubenswrapper[4907]: I0313 14:06:25.783246 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.880036 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:25 crc kubenswrapper[4907]: E0313 14:06:25.980763 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.081324 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.182324 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.282802 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.383859 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.484172 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.584689 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.684853 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.785798 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.886248 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:26 crc kubenswrapper[4907]: E0313 14:06:26.986944 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.088081 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.189170 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.289973 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.391020 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.491722 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.592345 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.692461 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.781478 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.783030 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.783171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:27 crc kubenswrapper[4907]: I0313 14:06:27.783197 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.792993 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.894064 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:27 crc kubenswrapper[4907]: E0313 14:06:27.994426 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.094985 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.195571 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.296159 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.396934 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.497341 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.598460 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.699406 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.799511 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:28 crc kubenswrapper[4907]: E0313 14:06:28.900016 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.000613 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.101693 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.202784 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.303481 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.404537 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.505447 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.605983 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.706798 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.782259 4907 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.783482 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.783529 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.783539 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:29 crc kubenswrapper[4907]: I0313 14:06:29.784152 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.784324 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.807316 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:29 crc kubenswrapper[4907]: E0313 14:06:29.908361 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.008731 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.109269 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.210094 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: I0313 14:06:30.242507 4907 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.310929 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.411961 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.512805 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.613477 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.714336 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.815461 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:30 crc kubenswrapper[4907]: E0313 14:06:30.916484 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.017589 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.118500 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.219423 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.320328 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.421346 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.521736 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.622008 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.723118 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.824133 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.875551 4907 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 14:06:31 crc kubenswrapper[4907]: E0313 14:06:31.924468 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.025478 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.126188 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.226920 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.327746 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.427945 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.528841 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.629911 4907 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.686265 4907 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732184 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732236 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.732263 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:32Z","lastTransitionTime":"2026-03-13T14:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.777817 4907 apiserver.go:52] "Watching apiserver" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.783177 4907 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.783585 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784136 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.784358 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784145 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.784407 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.784538 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.785489 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.785609 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786398 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786792 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.786923 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.787010 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.787068 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.787115 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.788290 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.790360 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.817491 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.819368 4907 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.831038 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837427 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837506 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837519 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837540 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.837552 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:32Z","lastTransitionTime":"2026-03-13T14:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.846022 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.858939 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.870287 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.880785 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.893851 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.907236 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917631 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917668 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917732 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917749 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917765 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917782 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917801 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917830 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917847 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917864 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917919 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917952 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917974 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.917991 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918267 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918815 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918861 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.918840 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919364 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919423 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919448 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919470 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919522 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919499 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919871 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.919917 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920137 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920137 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920155 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920165 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920173 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920301 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920352 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920439 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920730 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920741 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920737 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920769 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920826 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920925 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.920963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921589 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921941 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.921960 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922040 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922274 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922383 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922407 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922584 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922591 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922805 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922978 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.922938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923063 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923076 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923243 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923328 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923285 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923364 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923455 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923484 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923490 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923509 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923532 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923554 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923576 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923706 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923740 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923776 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923796 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923824 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924210 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924270 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923826 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.923967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925019 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925039 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925066 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925104 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925147 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925192 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925242 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924181 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925297 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925346 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925393 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924618 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.924685 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925451 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925503 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925581 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925453 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925618 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925491 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925673 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926063 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926115 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926469 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926555 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926609 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926661 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926711 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926762 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926858 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927049 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927110 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927209 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927254 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927327 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927376 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927425 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927512 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.925796 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.926995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927301 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927551 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927560 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927662 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927800 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927822 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927934 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927957 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927974 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.927999 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928033 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928057 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928152 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928202 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928242 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928281 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928315 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928351 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928436 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928593 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928635 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928706 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928735 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928799 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928831 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928877 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929002 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929079 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929117 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929149 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929219 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929250 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929280 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929336 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929487 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929535 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929572 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929609 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929642 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929677 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929712 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929748 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929784 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929945 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929979 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930045 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930077 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930112 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930146 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930222 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930258 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930299 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930332 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930400 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930505 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930538 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930571 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930604 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930637 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930669 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930808 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930844 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930969 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931005 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931079 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931122 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933255 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933946 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934034 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934246 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934321 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934439 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934482 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934522 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934561 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934600 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934859 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934919 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934943 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934963 4907 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934984 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935006 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935027 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935050 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935070 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935089 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935109 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935138 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935158 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935178 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935197 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935216 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935234 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935253 4907 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935275 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935294 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935312 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935330 4907 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935346 4907 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935364 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935383 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935405 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935424 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935441 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935462 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935480 4907 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935500 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935520 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935541 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935561 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935579 4907 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935599 4907 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935619 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935638 4907 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935656 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935673 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935691 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935712 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935731 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935752 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935770 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935790 4907 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935809 4907 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935828 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935847 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935867 4907 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935938 4907 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935957 4907 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935976 4907 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935996 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936929 4907 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.941998 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.441978151 +0000 UTC m=+92.341765840 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942184 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928230 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942207 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928551 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.928910 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942955 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.944003 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.943000 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.944700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945106 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.943218 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945632 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945929 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.945988 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946174 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946255 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929415 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930138 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930459 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930779 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947483 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930811 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930809 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947524 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.930982 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931289 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931482 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931563 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931533 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947588 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931643 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.931833 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932204 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932592 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932591 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932678 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.932922 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947976 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933423 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.948080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933611 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933945 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.933980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.934660 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935255 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935465 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935380 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935534 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.935582 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936114 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936373 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.936936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.937094 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.937297 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.938290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.938545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.938650 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939502 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939519 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939766 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.939813 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940045 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940350 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940632 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.940929 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941024 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941062 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941336 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941594 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941828 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941839 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.941856 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942237 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942251 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942397 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.948597 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.929008 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.946637 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.947358 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.942673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.948783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.949218 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.949535 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.950078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.950795 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964238 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.464202375 +0000 UTC m=+92.363990064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964426 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964493 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964527 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.950832 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.949759 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.957970 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964590 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964615 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.951036 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.951422 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.952158 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.952661 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.956679 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964567 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964677 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.964655 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:32Z","lastTransitionTime":"2026-03-13T14:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964707 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964728 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964766 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.46472099 +0000 UTC m=+92.364508679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.964802 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.464777672 +0000 UTC m=+92.364565371 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:32 crc kubenswrapper[4907]: E0313 14:06:32.965041 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:33.46502194 +0000 UTC m=+92.364809629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.965232 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.970512 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.971172 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.972479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.976743 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.977541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.977840 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982010 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982242 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982833 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982910 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.982957 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.983060 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.983726 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.983938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.984997 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.985692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986000 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986065 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986097 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.986230 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.987354 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.987934 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988594 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988710 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988798 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.988798 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989144 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989248 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.989301 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:32 crc kubenswrapper[4907]: I0313 14:06:32.992102 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.001379 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.001842 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.002524 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.015236 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037261 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037351 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037351 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037408 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037422 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037431 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037441 4907 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037451 4907 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037470 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037481 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037490 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037499 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037508 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037516 4907 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037525 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037535 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037543 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037551 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037562 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037573 4907 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037585 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037596 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037606 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037614 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037622 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037630 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037638 4907 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037646 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037655 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037664 4907 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037672 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037680 4907 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037688 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037697 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037705 4907 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037760 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037772 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037783 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037793 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037804 4907 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037817 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037829 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037839 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037850 4907 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037859 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037869 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037897 4907 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037907 4907 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037916 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037927 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037937 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037946 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037955 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037964 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037973 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037982 4907 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.037992 4907 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038002 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038011 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038020 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038029 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038039 4907 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038048 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038057 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038067 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038076 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038088 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038098 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038109 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038120 4907 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038130 4907 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038141 4907 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038150 4907 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038188 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038199 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038208 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038221 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038231 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038240 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038249 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038258 4907 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038268 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038277 4907 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038286 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038307 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038317 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038326 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038335 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038343 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038352 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038362 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038370 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038379 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038388 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038399 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038409 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038421 4907 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038431 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038440 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038450 4907 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038470 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038479 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038489 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038498 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038508 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038520 4907 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038530 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038539 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038548 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038557 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038568 4907 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038576 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038584 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038595 4907 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038604 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038614 4907 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038623 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038632 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038641 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038650 4907 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038659 4907 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038667 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038676 4907 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038685 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038694 4907 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038704 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038715 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038724 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038734 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038742 4907 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038752 4907 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038761 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038770 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038779 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038787 4907 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038796 4907 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038805 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038816 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038824 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.038833 4907 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.066988 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067031 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067060 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.067074 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.110584 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.125416 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.125823 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: source /etc/kubernetes/apiserver-url.env Mar 13 14:06:33 crc kubenswrapper[4907]: else Mar 13 14:06:33 crc kubenswrapper[4907]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 13 14:06:33 crc kubenswrapper[4907]: exit 1 Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.127247 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.137544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.138410 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 13 14:06:33 crc kubenswrapper[4907]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 13 14:06:33 crc kubenswrapper[4907]: ho_enable="--enable-hybrid-overlay" Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 13 14:06:33 crc kubenswrapper[4907]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 13 14:06:33 crc kubenswrapper[4907]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-host=127.0.0.1 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-port=9743 \ Mar 13 14:06:33 crc kubenswrapper[4907]: ${ho_enable} \ Mar 13 14:06:33 crc kubenswrapper[4907]: --enable-interconnect \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-approver \ Mar 13 14:06:33 crc kubenswrapper[4907]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --wait-for-kubernetes-api=200s \ Mar 13 14:06:33 crc kubenswrapper[4907]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.141144 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-webhook \ Mar 13 14:06:33 crc kubenswrapper[4907]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.142359 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 13 14:06:33 crc kubenswrapper[4907]: W0313 14:06:33.144872 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c WatchSource:0}: Error finding container 53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c: Status 404 returned error can't find the container with id 53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.147009 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.148185 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.163689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"53d2ba13c3bb87827b87557f503d37f81ee3023e9e9df3f8373171a31ea9559c"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.165185 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.165303 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"21c666661c2637edf91b1da9c29671bdf2ed4f25e257c5e4d8f5b72db45fdf0b"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.166348 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.166587 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"37ca6097afd5a084f52dfd5153bb04aced74584bac8e714df3e5dd76919c5caa"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.167217 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 13 14:06:33 crc kubenswrapper[4907]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 13 14:06:33 crc kubenswrapper[4907]: ho_enable="--enable-hybrid-overlay" Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 13 14:06:33 crc kubenswrapper[4907]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 13 14:06:33 crc kubenswrapper[4907]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-host=127.0.0.1 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --webhook-port=9743 \ Mar 13 14:06:33 crc kubenswrapper[4907]: ${ho_enable} \ Mar 13 14:06:33 crc kubenswrapper[4907]: --enable-interconnect \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-approver \ Mar 13 14:06:33 crc kubenswrapper[4907]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --wait-for-kubernetes-api=200s \ Mar 13 14:06:33 crc kubenswrapper[4907]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.168027 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: source /etc/kubernetes/apiserver-url.env Mar 13 14:06:33 crc kubenswrapper[4907]: else Mar 13 14:06:33 crc kubenswrapper[4907]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 13 14:06:33 crc kubenswrapper[4907]: exit 1 Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.169182 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.169253 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:06:33 crc kubenswrapper[4907]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 14:06:33 crc kubenswrapper[4907]: if [[ -f "/env/_master" ]]; then Mar 13 14:06:33 crc kubenswrapper[4907]: set -o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: source "/env/_master" Mar 13 14:06:33 crc kubenswrapper[4907]: set +o allexport Mar 13 14:06:33 crc kubenswrapper[4907]: fi Mar 13 14:06:33 crc kubenswrapper[4907]: Mar 13 14:06:33 crc kubenswrapper[4907]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 13 14:06:33 crc kubenswrapper[4907]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 14:06:33 crc kubenswrapper[4907]: --disable-webhook \ Mar 13 14:06:33 crc kubenswrapper[4907]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 13 14:06:33 crc kubenswrapper[4907]: --loglevel="${LOGLEVEL}" Mar 13 14:06:33 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 14:06:33 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170044 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170075 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170087 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170103 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.170117 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.170490 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.175209 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.190147 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.202917 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.216466 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.227339 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.240253 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.254548 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.265138 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272120 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272245 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272314 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272421 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.272510 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.282246 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.293629 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.303212 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.311868 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.374911 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.374975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.375000 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.375029 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.375048 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.443279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.443494 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.443455948 +0000 UTC m=+93.343243667 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478106 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478158 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478176 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478196 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.478212 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544711 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544854 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.544924 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.544959 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545055 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545026103 +0000 UTC m=+93.444813822 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545118 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545144 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545163 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545239 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545213198 +0000 UTC m=+93.445000917 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545326 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545374 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545361663 +0000 UTC m=+93.445149382 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545451 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545471 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545486 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: E0313 14:06:33.545528 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:34.545514568 +0000 UTC m=+93.445302287 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.580963 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581037 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581055 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581087 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.581115 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684003 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684072 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684092 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684119 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.684138 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787582 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787659 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.787672 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.788355 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.789592 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.791748 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.793042 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.794819 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.795405 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.796077 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.797575 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.799033 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.799845 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.800525 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.801913 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.802564 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.803741 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.804435 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.805578 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.806332 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.806824 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.808031 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.808762 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.809368 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.810633 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.811223 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.812494 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.813097 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.814189 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.814793 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.815254 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.816179 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.816642 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.817419 4907 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.817514 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.819077 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.820026 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.820394 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.821788 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.822995 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.823497 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.824495 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.825124 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.825543 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.826464 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.827377 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.828044 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.828910 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.829406 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.830272 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.830975 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.831735 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.832225 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.832659 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.833542 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.834078 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.834904 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.890265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.890537 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.890805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.891041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.891185 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993908 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993950 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993963 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.993981 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:33 crc kubenswrapper[4907]: I0313 14:06:33.994026 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:33Z","lastTransitionTime":"2026-03-13T14:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096008 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096062 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096079 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096103 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.096120 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198661 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198714 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198730 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.198740 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.301279 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.301592 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.301840 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.302105 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.302311 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.405694 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.405997 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.406081 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.406150 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.406221 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.453362 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.453747 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.453711167 +0000 UTC m=+95.353498896 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.509865 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.509957 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.509975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.510003 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.510021 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.554979 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.555283 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555277 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.555455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555589 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.55555414 +0000 UTC m=+95.455341859 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555395 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555679 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555742 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.555851 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.555831238 +0000 UTC m=+95.455618957 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.555987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556134 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556258 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.55622818 +0000 UTC m=+95.456015909 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556384 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556417 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556437 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.556496 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:36.556471568 +0000 UTC m=+95.456259287 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613508 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613561 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613580 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613605 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.613622 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.717708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718687 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718798 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.718824 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.782391 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.782480 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.782391 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.782628 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.782796 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:34 crc kubenswrapper[4907]: E0313 14:06:34.783001 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821462 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821505 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821531 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.821544 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924266 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924333 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924353 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:34 crc kubenswrapper[4907]: I0313 14:06:34.924408 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:34Z","lastTransitionTime":"2026-03-13T14:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027495 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027540 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.027558 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130529 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130567 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130590 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.130601 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232673 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232731 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232747 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.232757 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336073 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336085 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336102 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.336112 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439123 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439208 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.439311 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544061 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544119 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544139 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544164 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.544182 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647757 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.647774 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751696 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751794 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751815 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751848 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.751871 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855667 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855712 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855722 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855743 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.855755 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958723 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958788 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958833 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:35 crc kubenswrapper[4907]: I0313 14:06:35.958855 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:35Z","lastTransitionTime":"2026-03-13T14:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010343 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010359 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010380 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.010392 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.020190 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023440 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023491 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023508 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.023519 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.037836 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042318 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042458 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.042838 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.058143 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062597 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062644 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062664 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.062677 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.078289 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.082863 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.082958 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.082976 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.083368 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.083421 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.099053 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.099303 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103216 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103303 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103357 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103446 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.103507 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.205935 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206230 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.206628 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310289 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.310322 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413334 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413352 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413379 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.413398 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.476281 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.476485 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.476461712 +0000 UTC m=+99.376249401 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516693 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516720 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.516742 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.576991 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.577049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.577086 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.577117 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577226 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577243 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577262 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577276 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577283 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577263753 +0000 UTC m=+99.477051442 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577281 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577307 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577297684 +0000 UTC m=+99.477085373 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577324 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577350 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577352 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577385 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577375608 +0000 UTC m=+99.477163297 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.577421 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:40.577394898 +0000 UTC m=+99.477182617 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620179 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620250 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.620286 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723077 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723111 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723141 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.723153 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.781848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.781929 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.782074 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.782089 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.782250 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:36 crc kubenswrapper[4907]: E0313 14:06:36.782349 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826000 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826059 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826076 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826101 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.826120 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929321 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929393 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929411 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929435 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:36 crc kubenswrapper[4907]: I0313 14:06:36.929452 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:36Z","lastTransitionTime":"2026-03-13T14:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032149 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032256 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032281 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.032297 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136416 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136478 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136495 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136520 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.136538 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.239975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240037 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240053 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240077 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.240094 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343165 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343259 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343619 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343669 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.343697 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447337 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447354 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447379 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.447398 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550553 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550602 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550612 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.550645 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.658622 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.658978 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.659064 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.659148 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.659223 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762169 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762232 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762250 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762276 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.762294 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.864532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865212 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865331 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.865596 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967301 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967336 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967345 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967363 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:37 crc kubenswrapper[4907]: I0313 14:06:37.967374 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:37Z","lastTransitionTime":"2026-03-13T14:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069322 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069384 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.069471 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171439 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171450 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171466 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.171478 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273301 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273310 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273322 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.273332 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.376486 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.376857 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.377084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.377243 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.377372 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.479931 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.479974 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.479989 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.480011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.480029 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583139 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583188 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583200 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583222 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.583234 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686716 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686784 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.686823 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.781343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.781456 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.781491 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:38 crc kubenswrapper[4907]: E0313 14:06:38.781647 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:38 crc kubenswrapper[4907]: E0313 14:06:38.781812 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:38 crc kubenswrapper[4907]: E0313 14:06:38.781984 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789803 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789871 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789938 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.789975 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.790001 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893748 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893785 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893794 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893808 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.893818 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996919 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996934 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996955 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:38 crc kubenswrapper[4907]: I0313 14:06:38.996969 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:38Z","lastTransitionTime":"2026-03-13T14:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.099961 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100037 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100056 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.100104 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202411 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202492 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202515 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202547 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.202570 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306156 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306241 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.306319 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409048 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409109 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409131 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.409146 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512481 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512564 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512593 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.512632 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.615822 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.615916 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.615980 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.616007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.616027 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.718930 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.718977 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.718993 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.719018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.719038 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821512 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821568 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821584 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821609 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.821628 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924734 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924852 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924918 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:39 crc kubenswrapper[4907]: I0313 14:06:39.924936 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:39Z","lastTransitionTime":"2026-03-13T14:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027268 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027326 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027345 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027369 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.027385 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130487 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130498 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130516 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.130528 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.233396 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.233540 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.234001 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.234081 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.234110 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337178 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337270 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337304 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337337 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.337520 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440564 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440627 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440644 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.440685 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.516271 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.516567 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.516527977 +0000 UTC m=+107.416315706 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544355 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544371 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544402 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.544474 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617555 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617660 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.617705 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.617758 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618014 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618054 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618232 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618101 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.618078581 +0000 UTC m=+107.517866300 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618457 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.618436433 +0000 UTC m=+107.518224152 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.617964 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618551 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618578 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618595 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618679 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.618639509 +0000 UTC m=+107.518427228 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.618707 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:06:48.61869281 +0000 UTC m=+107.518480539 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648432 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648472 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648496 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648514 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.648526 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.750868 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.750989 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.751015 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.751046 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.751069 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.782155 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.782319 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.782414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.782556 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.783980 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.782868 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.800966 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.801235 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:40 crc kubenswrapper[4907]: E0313 14:06:40.801556 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855688 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855733 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855748 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.855784 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958610 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958679 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958701 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:40 crc kubenswrapper[4907]: I0313 14:06:40.958714 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:40Z","lastTransitionTime":"2026-03-13T14:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061317 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061347 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061371 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.061380 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.147441 4907 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164616 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.164661 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.187182 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:41 crc kubenswrapper[4907]: E0313 14:06:41.187750 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267187 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267262 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267279 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267309 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.267342 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370336 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370399 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370408 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370425 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.370435 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473791 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.473835 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576810 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576938 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.576981 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.577003 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680751 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680771 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680801 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.680818 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783199 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783233 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783241 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783254 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.783265 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.799455 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.808448 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.823976 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.835202 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.847060 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.862797 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.880529 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885477 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885525 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885543 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885568 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.885622 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988695 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988761 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988778 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988829 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:41 crc kubenswrapper[4907]: I0313 14:06:41.988848 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:41Z","lastTransitionTime":"2026-03-13T14:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092650 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092729 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092757 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.092780 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195449 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195517 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.195592 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.297866 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298002 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298046 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.298064 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.400965 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401017 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401034 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401060 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.401078 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506228 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506298 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506316 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506342 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.506360 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609363 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609430 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609447 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609476 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.609493 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713188 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713298 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713331 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.713355 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.782391 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:42 crc kubenswrapper[4907]: E0313 14:06:42.782576 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.782682 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:42 crc kubenswrapper[4907]: E0313 14:06:42.782816 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.782390 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:42 crc kubenswrapper[4907]: E0313 14:06:42.783040 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818790 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818932 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818962 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.818999 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.819025 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922294 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922366 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922384 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922412 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:42 crc kubenswrapper[4907]: I0313 14:06:42.922431 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:42Z","lastTransitionTime":"2026-03-13T14:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.026925 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.026996 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.027016 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.027044 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.027106 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130044 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130086 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130099 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130119 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.130132 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233338 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.233405 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.335933 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336015 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336065 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.336084 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439198 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439269 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439287 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439366 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.439393 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543454 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543547 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543608 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.543704 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647448 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647552 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647584 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.647607 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751834 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751856 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751957 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.751981 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855398 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855483 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855508 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.855566 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959374 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959442 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959460 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959489 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:43 crc kubenswrapper[4907]: I0313 14:06:43.959508 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:43Z","lastTransitionTime":"2026-03-13T14:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062715 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062738 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.062842 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166447 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166511 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166562 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.166588 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270129 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270263 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.270319 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373494 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373578 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373607 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373646 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.373671 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476731 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476829 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476857 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.476918 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580185 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580292 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580325 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.580351 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.682901 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.682969 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.682981 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.683004 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.683017 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.782241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.782379 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:44 crc kubenswrapper[4907]: E0313 14:06:44.782434 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.782481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:44 crc kubenswrapper[4907]: E0313 14:06:44.782649 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:44 crc kubenswrapper[4907]: E0313 14:06:44.782983 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785614 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785651 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785662 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785680 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.785693 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889042 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889088 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889099 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889117 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.889133 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993707 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993777 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993796 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993833 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:44 crc kubenswrapper[4907]: I0313 14:06:44.993856 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:44Z","lastTransitionTime":"2026-03-13T14:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097290 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097358 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097378 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.097425 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199858 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199913 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199978 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.199999 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.202232 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.202419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.220613 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.255436 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.279865 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.299046 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309542 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309553 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309569 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.309580 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.321677 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.337093 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.349826 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414343 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414412 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414432 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414464 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.414482 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.517854 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518243 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518437 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.518929 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622624 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.622688 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725860 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725950 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725965 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.725987 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.726003 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.827952 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828186 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828389 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.828476 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.931671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.931954 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.932016 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.932085 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:45 crc kubenswrapper[4907]: I0313 14:06:45.932143 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:45Z","lastTransitionTime":"2026-03-13T14:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034468 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034751 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034810 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.034863 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.137740 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138375 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138524 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.138732 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238040 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238115 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238138 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238165 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.238184 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.262399 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268181 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268255 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268276 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268305 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.268325 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.290953 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.297268 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.297510 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.298147 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.298559 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.298816 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.322585 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.331437 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.331709 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.332601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.333326 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.333584 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.357546 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.364683 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.365100 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.365492 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.366159 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.366394 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.389841 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:46Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.391157 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394594 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394640 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.394658 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498306 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498384 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498414 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498473 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.498495 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.601818 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602175 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602278 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602380 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.602515 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.706526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.706973 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.707059 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.707249 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.707372 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.782354 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.782408 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.782355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.782560 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.782707 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:46 crc kubenswrapper[4907]: E0313 14:06:46.782871 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810940 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810964 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.810992 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.811012 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914355 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914441 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914464 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914490 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:46 crc kubenswrapper[4907]: I0313 14:06:46.914512 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:46Z","lastTransitionTime":"2026-03-13T14:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.018651 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019070 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019357 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.019486 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123679 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123758 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123780 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123808 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.123828 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.210370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227364 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227443 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227550 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.227634 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.248275 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.268284 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.286511 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.304511 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.322779 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-l7zg7"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.323363 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.327298 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.327663 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.327814 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.328648 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.332566 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.332763 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.332936 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.333071 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.333186 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.348319 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.368219 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.385037 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.400222 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.414188 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435739 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435777 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435786 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.435814 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.437420 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.454373 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.468939 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.483229 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.490927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szmh9\" (UniqueName: \"kubernetes.io/projected/d105034b-f041-4fb5-8df6-6b1bb7681fda-kube-api-access-szmh9\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.491194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d105034b-f041-4fb5-8df6-6b1bb7681fda-hosts-file\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.538990 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539238 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539354 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539543 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.539668 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.592554 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d105034b-f041-4fb5-8df6-6b1bb7681fda-hosts-file\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.592662 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szmh9\" (UniqueName: \"kubernetes.io/projected/d105034b-f041-4fb5-8df6-6b1bb7681fda-kube-api-access-szmh9\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.593178 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d105034b-f041-4fb5-8df6-6b1bb7681fda-hosts-file\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.617820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szmh9\" (UniqueName: \"kubernetes.io/projected/d105034b-f041-4fb5-8df6-6b1bb7681fda-kube-api-access-szmh9\") pod \"node-resolver-l7zg7\" (UID: \"d105034b-f041-4fb5-8df6-6b1bb7681fda\") " pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642549 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642773 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642942 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.642680 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-l7zg7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.643264 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.643338 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: W0313 14:06:47.656307 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd105034b_f041_4fb5_8df6_6b1bb7681fda.slice/crio-43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db WatchSource:0}: Error finding container 43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db: Status 404 returned error can't find the container with id 43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.697857 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xtwf7"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.698152 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-p8ptc"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.698782 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.699165 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.700245 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-hm56j"] Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.700535 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.704738 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.704793 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.704819 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705016 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705075 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705099 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.705224 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706315 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706509 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706748 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.706865 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.708315 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.716967 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.728990 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.740751 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749605 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749652 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749666 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749689 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.749707 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.759709 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.769658 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.780829 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794440 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-conf-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794504 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3fb4213-144f-4b01-9d78-ff3315761bb3-mcd-auth-proxy-config\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxwbn\" (UniqueName: \"kubernetes.io/projected/f3fb4213-144f-4b01-9d78-ff3315761bb3-kube-api-access-vxwbn\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794603 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3fb4213-144f-4b01-9d78-ff3315761bb3-proxy-tls\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-k8s-cni-cncf-io\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794664 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-kubelet\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.794835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f3fb4213-144f-4b01-9d78-ff3315761bb3-rootfs\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795025 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795042 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-bin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795125 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-multus\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-hostroot\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795183 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-socket-dir-parent\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795197 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-multus-certs\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795312 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-system-cni-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795384 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-cnibin\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795413 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vldr\" (UniqueName: \"kubernetes.io/projected/57e4e912-9a98-4419-87ce-f16e178ed921-kube-api-access-7vldr\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795437 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-multus-daemon-config\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795462 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-cnibin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795480 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-netns\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-etc-kubernetes\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795519 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-os-release\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cwbq\" (UniqueName: \"kubernetes.io/projected/64ebd338-0158-44dc-90ab-e7c285e87762-kube-api-access-6cwbq\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795568 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-cni-binary-copy\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795610 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-os-release\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.795633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-system-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.796226 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.814734 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.826657 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.836207 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.849001 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854673 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854682 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854707 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.854718 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.863835 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.879378 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896495 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-socket-dir-parent\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896548 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-multus-certs\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896572 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-system-cni-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-cnibin\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896616 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vldr\" (UniqueName: \"kubernetes.io/projected/57e4e912-9a98-4419-87ce-f16e178ed921-kube-api-access-7vldr\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896638 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-multus-daemon-config\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896667 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-netns\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-etc-kubernetes\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-cnibin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896727 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-os-release\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896748 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cwbq\" (UniqueName: \"kubernetes.io/projected/64ebd338-0158-44dc-90ab-e7c285e87762-kube-api-access-6cwbq\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896770 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-cni-binary-copy\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896816 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-os-release\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-system-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-conf-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896897 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3fb4213-144f-4b01-9d78-ff3315761bb3-mcd-auth-proxy-config\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896913 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxwbn\" (UniqueName: \"kubernetes.io/projected/f3fb4213-144f-4b01-9d78-ff3315761bb3-kube-api-access-vxwbn\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896951 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.896968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3fb4213-144f-4b01-9d78-ff3315761bb3-proxy-tls\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897002 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-kubelet\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f3fb4213-144f-4b01-9d78-ff3315761bb3-rootfs\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-k8s-cni-cncf-io\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-bin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-multus\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-hostroot\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897253 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897304 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897645 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-socket-dir-parent\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-os-release\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897725 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-multus-certs\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897759 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-system-cni-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897789 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-cnibin\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-system-cni-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897901 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-cni-binary-copy\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-multus-conf-dir\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.897979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-netns\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898009 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-etc-kubernetes\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898038 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-kubelet\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898049 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-cnibin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898079 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-os-release\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898129 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/64ebd338-0158-44dc-90ab-e7c285e87762-multus-daemon-config\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-hostroot\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898572 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-multus\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f3fb4213-144f-4b01-9d78-ff3315761bb3-rootfs\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898699 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/57e4e912-9a98-4419-87ce-f16e178ed921-tuning-conf-dir\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-run-k8s-cni-cncf-io\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898769 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3fb4213-144f-4b01-9d78-ff3315761bb3-mcd-auth-proxy-config\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.899171 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.898247 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64ebd338-0158-44dc-90ab-e7c285e87762-host-var-lib-cni-bin\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.901671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f3fb4213-144f-4b01-9d78-ff3315761bb3-proxy-tls\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.908602 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/57e4e912-9a98-4419-87ce-f16e178ed921-cni-binary-copy\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.913172 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.915857 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cwbq\" (UniqueName: \"kubernetes.io/projected/64ebd338-0158-44dc-90ab-e7c285e87762-kube-api-access-6cwbq\") pod \"multus-xtwf7\" (UID: \"64ebd338-0158-44dc-90ab-e7c285e87762\") " pod="openshift-multus/multus-xtwf7" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.916348 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vldr\" (UniqueName: \"kubernetes.io/projected/57e4e912-9a98-4419-87ce-f16e178ed921-kube-api-access-7vldr\") pod \"multus-additional-cni-plugins-p8ptc\" (UID: \"57e4e912-9a98-4419-87ce-f16e178ed921\") " pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.917782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxwbn\" (UniqueName: \"kubernetes.io/projected/f3fb4213-144f-4b01-9d78-ff3315761bb3-kube-api-access-vxwbn\") pod \"machine-config-daemon-hm56j\" (UID: \"f3fb4213-144f-4b01-9d78-ff3315761bb3\") " pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.927869 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.940868 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.955201 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.956749 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957176 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957249 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957327 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.957409 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:47Z","lastTransitionTime":"2026-03-13T14:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.968555 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.981577 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:47 crc kubenswrapper[4907]: I0313 14:06:47.993473 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:47Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.019032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.030256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xtwf7" Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.039455 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57e4e912_9a98_4419_87ce_f16e178ed921.slice/crio-079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095 WatchSource:0}: Error finding container 079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095: Status 404 returned error can't find the container with id 079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095 Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.040664 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.047580 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64ebd338_0158_44dc_90ab_e7c285e87762.slice/crio-f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3 WatchSource:0}: Error finding container f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3: Status 404 returned error can't find the container with id f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3 Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.059069 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3fb4213_144f_4b01_9d78_ff3315761bb3.slice/crio-7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e WatchSource:0}: Error finding container 7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e: Status 404 returned error can't find the container with id 7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059489 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059620 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059654 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.059664 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.067519 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.069254 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071269 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071430 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071692 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071825 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.071937 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.072074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.074492 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.083362 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.094025 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.101891 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.118565 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.132519 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.146615 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.160162 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161818 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161863 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161876 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161915 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.161931 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.174111 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.185295 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199136 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199505 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199736 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199793 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199815 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199846 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199867 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199898 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199916 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199932 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199973 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.199991 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200022 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200039 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200180 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200232 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200304 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.200328 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.210672 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.213977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.215584 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerStarted","Data":"079e22966c2667db80b9a401ae908cc814dd5201d0acb692a9e8a44a1136c095"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.217129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.217155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"f25904e76871efac8ccd4826678dfa90aa051973756ce2119ad33d1abda0b3b3"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.218493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l7zg7" event={"ID":"d105034b-f041-4fb5-8df6-6b1bb7681fda","Type":"ContainerStarted","Data":"21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.218524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-l7zg7" event={"ID":"d105034b-f041-4fb5-8df6-6b1bb7681fda","Type":"ContainerStarted","Data":"43541ca6821c8fc2f866fd8525a0781a2c00d4fd50c8c0b176be0d71affd01db"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.220155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.220220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"7bdf72caf166aa15007523a91a3e3eae10bae5574b7406aeebd6c8d935e0be6e"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.230811 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.245352 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.255778 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264539 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264580 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264588 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264606 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.264616 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.271361 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.287539 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300330 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300924 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300943 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300971 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.300987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301001 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301168 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301206 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301229 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301254 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301533 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301041 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301553 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301637 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301751 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301083 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301967 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301969 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301981 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301654 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.301129 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302043 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302018 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302040 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302072 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302063 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.302608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.304471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.315500 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.318820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"ovnkube-node-qxlx5\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.327087 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.338566 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.349340 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.364570 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367059 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367138 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367163 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.367181 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.378757 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.384634 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:48 crc kubenswrapper[4907]: W0313 14:06:48.397852 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1006376d_4033_454e_8a23_2b67882bdd0b.slice/crio-4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0 WatchSource:0}: Error finding container 4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0: Status 404 returned error can't find the container with id 4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0 Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.402663 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469770 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469782 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.469814 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572569 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572915 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572925 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572945 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.572955 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.604094 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.604282 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.604249229 +0000 UTC m=+123.504036918 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676320 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676330 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.676358 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705343 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.705365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705500 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705516 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705528 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705573 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.705557956 +0000 UTC m=+123.605345645 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705868 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705984 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706031 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706048 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706112 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.706089972 +0000 UTC m=+123.605877671 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.705929 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706168 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.706158684 +0000 UTC m=+123.605946393 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.706277 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.706252997 +0000 UTC m=+123.606040686 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779475 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779807 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.779914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.780009 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.781689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.781716 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.781689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.781800 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.781899 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:48 crc kubenswrapper[4907]: E0313 14:06:48.781967 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883421 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883432 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883448 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.883458 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.986740 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987068 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987165 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987433 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:48 crc kubenswrapper[4907]: I0313 14:06:48.987511 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:48Z","lastTransitionTime":"2026-03-13T14:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.091471 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.091827 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.092042 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.092342 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.092556 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195557 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195623 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195642 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.195687 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.227261 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f" exitCode=0 Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.227353 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.229266 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" exitCode=0 Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.229428 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.229565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.234525 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.253605 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.269094 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.281647 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.295309 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.297800 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.297992 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.298017 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.298033 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.298043 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.308662 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.320629 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.335975 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.347123 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.370295 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.385534 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402687 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402744 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.402790 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.406204 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.414622 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.424663 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.434075 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.444775 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.459774 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.494305 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505438 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505494 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505514 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.505525 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.521457 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.536763 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.552084 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.565365 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.581227 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.595182 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608308 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608352 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608365 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608387 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.608402 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.609545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:49Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710187 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710239 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710258 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.710271 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816194 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816599 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816610 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.816643 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919498 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919548 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919557 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919571 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:49 crc kubenswrapper[4907]: I0313 14:06:49.919584 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:49Z","lastTransitionTime":"2026-03-13T14:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022274 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022321 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022344 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022359 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.022372 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124850 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124900 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124912 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124929 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.124941 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227224 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227523 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227535 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227549 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.227558 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241499 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241516 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.241524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.243036 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302" exitCode=0 Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.243778 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.257184 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.273017 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.286598 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.307645 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.322683 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330682 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330691 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330705 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.330713 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.334055 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.346156 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.357993 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.368529 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.380633 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.391943 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.405468 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:50Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434475 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434523 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434534 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434550 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.434560 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536599 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536640 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536656 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.536666 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.638995 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639040 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639054 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639073 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.639091 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741310 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741328 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.741379 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.782407 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.782508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.782421 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:50 crc kubenswrapper[4907]: E0313 14:06:50.782623 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:50 crc kubenswrapper[4907]: E0313 14:06:50.782780 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:50 crc kubenswrapper[4907]: E0313 14:06:50.782911 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843799 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843831 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.843844 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.946965 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947033 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947052 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947079 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:50 crc kubenswrapper[4907]: I0313 14:06:50.947099 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:50Z","lastTransitionTime":"2026-03-13T14:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052876 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052931 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052959 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.052977 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156307 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156368 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156386 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156414 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.156432 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.250743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.254072 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45" exitCode=0 Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.254159 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260212 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260226 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260243 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.260255 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.270646 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.283827 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.302345 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.322962 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.343381 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.358931 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363635 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363687 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363706 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.363719 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.374264 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.389273 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.404785 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.422236 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.441427 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.458873 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468642 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468727 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468755 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.468772 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571136 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571147 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571166 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.571178 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673150 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673179 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673187 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673201 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.673209 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776284 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776339 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776377 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.776394 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.801712 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.817400 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.837495 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.860627 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.875078 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879545 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879626 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879655 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.879678 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.897813 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.915609 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.934358 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.950720 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.973145 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982765 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982823 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982868 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.982913 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:51Z","lastTransitionTime":"2026-03-13T14:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:51 crc kubenswrapper[4907]: I0313 14:06:51.988826 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.011410 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085586 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085664 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085689 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.085757 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188842 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188935 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188956 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.188984 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.189002 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.260104 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818" exitCode=0 Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.260156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.272726 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291722 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291569 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291752 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.291960 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.303943 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.321257 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.331990 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.345509 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.361131 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.372407 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.386176 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394557 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394603 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.394644 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.403808 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.414401 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.433192 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496519 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496813 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496821 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496835 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.496844 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598647 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598680 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598689 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598705 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.598715 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701768 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701812 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701824 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.701854 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.782324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.782343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:52 crc kubenswrapper[4907]: E0313 14:06:52.782538 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.782357 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:52 crc kubenswrapper[4907]: E0313 14:06:52.782739 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:52 crc kubenswrapper[4907]: E0313 14:06:52.782902 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805752 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805811 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805828 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805856 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.805875 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.908952 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909024 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909057 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:52 crc kubenswrapper[4907]: I0313 14:06:52.909075 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:52Z","lastTransitionTime":"2026-03-13T14:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012217 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012265 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012296 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.012307 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115488 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115537 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115554 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115577 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.115595 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218646 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218694 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218726 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.218739 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.275687 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847" exitCode=0 Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.275805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.284329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.293177 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.307039 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321541 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321597 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321615 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321641 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.321661 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.333700 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.354248 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.371571 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.399353 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438185 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438237 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438254 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438275 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.438290 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.450953 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.467980 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.483327 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.498139 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.510611 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.539864 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:53Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.543970 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.543991 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.543999 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.544013 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.544021 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648067 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648076 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648102 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.648116 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751405 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751535 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751598 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751630 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.751682 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.854843 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855362 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855469 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.855553 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.958801 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959457 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959518 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:53 crc kubenswrapper[4907]: I0313 14:06:53.959538 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:53Z","lastTransitionTime":"2026-03-13T14:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063743 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063824 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063846 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063873 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.063939 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.103917 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-745zx"] Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.104507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.108826 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.109029 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.109241 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.109452 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.128351 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.147359 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.163856 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167277 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167447 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167532 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167631 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.167717 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.176276 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58e11265-2fd9-4b9b-926c-3343bb84ee53-serviceca\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.176387 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58e11265-2fd9-4b9b-926c-3343bb84ee53-host\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.176453 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnz7b\" (UniqueName: \"kubernetes.io/projected/58e11265-2fd9-4b9b-926c-3343bb84ee53-kube-api-access-mnz7b\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.195427 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.214833 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.236525 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.252475 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.270537 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271100 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271214 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271299 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271395 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.271493 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.277645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58e11265-2fd9-4b9b-926c-3343bb84ee53-serviceca\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.277737 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58e11265-2fd9-4b9b-926c-3343bb84ee53-host\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.277787 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnz7b\" (UniqueName: \"kubernetes.io/projected/58e11265-2fd9-4b9b-926c-3343bb84ee53-kube-api-access-mnz7b\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.278377 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58e11265-2fd9-4b9b-926c-3343bb84ee53-host\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.279624 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58e11265-2fd9-4b9b-926c-3343bb84ee53-serviceca\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.291844 4907 generic.go:334] "Generic (PLEG): container finished" podID="57e4e912-9a98-4419-87ce-f16e178ed921" containerID="4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120" exitCode=0 Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.291944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerDied","Data":"4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.297693 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.306733 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnz7b\" (UniqueName: \"kubernetes.io/projected/58e11265-2fd9-4b9b-926c-3343bb84ee53-kube-api-access-mnz7b\") pod \"node-ca-745zx\" (UID: \"58e11265-2fd9-4b9b-926c-3343bb84ee53\") " pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.316148 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.337385 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.358692 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375748 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375839 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375868 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.375913 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.377288 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.395435 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.412244 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.431263 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.432788 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-745zx" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.456291 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: W0313 14:06:54.462605 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58e11265_2fd9_4b9b_926c_3343bb84ee53.slice/crio-f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad WatchSource:0}: Error finding container f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad: Status 404 returned error can't find the container with id f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.473662 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479072 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479120 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479137 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479163 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.479182 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.487372 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.503502 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.524685 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.544549 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.572376 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581201 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581241 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581271 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.581286 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.591159 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.606368 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.620418 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:54Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683677 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683720 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683771 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.683807 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.781983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.782028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.782109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:54 crc kubenswrapper[4907]: E0313 14:06:54.782408 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786454 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786488 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786499 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786516 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.786529 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: E0313 14:06:54.787124 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:54 crc kubenswrapper[4907]: E0313 14:06:54.787403 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889246 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889284 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889315 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.889326 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991449 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991515 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991533 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991562 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:54 crc kubenswrapper[4907]: I0313 14:06:54.991580 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:54Z","lastTransitionTime":"2026-03-13T14:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096205 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096275 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096324 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.096343 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199139 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199199 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199212 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199232 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.199245 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.301098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" event={"ID":"57e4e912-9a98-4419-87ce-f16e178ed921","Type":"ContainerStarted","Data":"38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304200 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304285 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-745zx" event={"ID":"58e11265-2fd9-4b9b-926c-3343bb84ee53","Type":"ContainerStarted","Data":"4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-745zx" event={"ID":"58e11265-2fd9-4b9b-926c-3343bb84ee53","Type":"ContainerStarted","Data":"f83bd19b71e3bdfc0d53ed197f3b3f54ce16f8b3d4028e9e8b9b0cc846eba6ad"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304311 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304861 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.304957 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.314974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.315384 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.315409 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.329564 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.343641 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.350384 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.366753 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.394552 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408466 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408543 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408581 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408607 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.408619 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.413727 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.435446 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.446855 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.462324 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.476456 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.491458 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.507460 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511668 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511713 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511727 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511750 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.511762 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.521418 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.534419 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.549055 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.568008 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.587471 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.601639 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.613643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615366 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615402 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615414 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615434 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.615453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.629207 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.644678 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.659258 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.686266 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.699764 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.711008 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718643 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718715 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718731 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718752 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.718766 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.731614 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.744542 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:55Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.783143 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823586 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823616 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823624 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.823648 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926658 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926690 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926699 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926714 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:55 crc kubenswrapper[4907]: I0313 14:06:55.926726 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:55Z","lastTransitionTime":"2026-03-13T14:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030598 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030639 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030653 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.030681 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.134911 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.134988 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.135001 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.135023 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.135041 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.237924 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.237993 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.238007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.238031 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.238044 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.331650 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.335490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.336834 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341716 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341786 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341816 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341848 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.341872 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.363657 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.380371 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.383643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.400364 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.419744 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.436029 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444719 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444785 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444804 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444830 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.444848 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.449598 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.472198 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.487984 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.502806 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.514594 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.529380 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.545531 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547697 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.547824 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.562479 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.579832 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.603415 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.629617 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.642212 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651795 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651843 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651855 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651898 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.651912 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.660671 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.678826 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.694367 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.706812 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.720104 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.734824 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.752545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754645 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754669 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.754687 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.765018 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781500 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.781650 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781739 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.781962 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.782030 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.782100 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790242 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790260 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.790276 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.802359 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806368 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806407 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806419 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806437 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.806451 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.822399 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.830946 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.830987 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.830999 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.831018 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.831030 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.841614 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845192 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845224 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845252 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.845262 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.859001 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862315 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862339 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862347 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862361 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.862370 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.873302 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:56Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:56 crc kubenswrapper[4907]: E0313 14:06:56.873420 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874704 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874733 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874745 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.874755 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976705 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976764 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976781 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976806 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:56 crc kubenswrapper[4907]: I0313 14:06:56.976824 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:56Z","lastTransitionTime":"2026-03-13T14:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079375 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079449 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079467 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079495 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.079518 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183422 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183494 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183513 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183548 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.183566 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287299 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287357 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287374 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287429 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.287453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390823 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390904 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390916 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390938 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.390951 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494017 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494069 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494083 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494106 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.494121 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596617 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596660 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596671 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596690 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.596702 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.699688 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.699941 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.700024 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.700098 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.700186 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802791 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802834 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802849 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802871 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.802927 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905135 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905603 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905614 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905628 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:57 crc kubenswrapper[4907]: I0313 14:06:57.905636 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:57Z","lastTransitionTime":"2026-03-13T14:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008054 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008091 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008101 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008140 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.008160 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111040 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111080 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111091 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111110 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.111122 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214462 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214526 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214544 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.214592 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317309 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317392 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317411 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317439 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.317456 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.345702 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/0.log" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.350322 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7" exitCode=1 Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.350362 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.351783 4907 scope.go:117] "RemoveContainer" containerID="569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.374607 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.392628 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.415725 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423288 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423320 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423329 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423342 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.423351 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.429303 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.449696 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:57Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.947520 6737 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.946905 6737 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 14:06:57.949275 6737 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:06:57.949375 6737 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:06:57.949412 6737 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 14:06:57.949440 6737 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0313 14:06:57.949483 6737 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0313 14:06:57.949481 6737 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:06:57.949508 6737 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0313 14:06:57.949515 6737 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:06:57.949523 6737 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:06:57.949543 6737 factory.go:656] Stopping watch factory\\\\nI0313 14:06:57.949567 6737 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.466690 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.488030 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.503076 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.515510 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.524740 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526177 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526228 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526240 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526256 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.526265 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.540655 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.552621 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.568144 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629253 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629301 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629315 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629338 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.629356 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733141 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733209 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733322 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733354 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.733403 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.782028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.782136 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.782049 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:06:58 crc kubenswrapper[4907]: E0313 14:06:58.782257 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:06:58 crc kubenswrapper[4907]: E0313 14:06:58.782475 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:06:58 crc kubenswrapper[4907]: E0313 14:06:58.782597 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836630 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836700 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836722 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836753 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.836773 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939331 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939373 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939382 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939396 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:58 crc kubenswrapper[4907]: I0313 14:06:58.939404 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:58Z","lastTransitionTime":"2026-03-13T14:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042613 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042667 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042685 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042711 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.042730 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144823 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144871 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144914 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144936 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.144953 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247210 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247260 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247291 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.247303 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.349993 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350074 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350091 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.350100 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.354642 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.355386 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/0.log" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.357922 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" exitCode=1 Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.357982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.358021 4907 scope.go:117] "RemoveContainer" containerID="569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.358726 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:06:59 crc kubenswrapper[4907]: E0313 14:06:59.358956 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.371836 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.399490 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:57Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.947520 6737 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.946905 6737 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 14:06:57.949275 6737 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:06:57.949375 6737 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:06:57.949412 6737 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 14:06:57.949440 6737 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0313 14:06:57.949483 6737 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0313 14:06:57.949481 6737 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:06:57.949508 6737 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0313 14:06:57.949515 6737 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:06:57.949523 6737 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:06:57.949543 6737 factory.go:656] Stopping watch factory\\\\nI0313 14:06:57.949567 6737 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.412066 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.427522 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.440603 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455050 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455084 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455092 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455107 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.455118 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.467867 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.481604 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.496925 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.510637 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.522188 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.535023 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.544501 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.554106 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:06:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557826 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557865 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557893 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557911 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.557924 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.660945 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661025 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661050 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661085 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.661140 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764732 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764798 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764819 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764844 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.764916 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.866936 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.866994 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.867011 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.867036 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.867056 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.969383 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.969718 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.969841 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.970009 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:06:59 crc kubenswrapper[4907]: I0313 14:06:59.970132 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:06:59Z","lastTransitionTime":"2026-03-13T14:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.019302 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4"] Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.020223 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.023117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.023182 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.038063 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050278 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050440 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050494 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.050520 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bg45\" (UniqueName: \"kubernetes.io/projected/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-kube-api-access-6bg45\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.061057 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072615 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072655 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072666 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072683 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.072694 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.073453 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.088154 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.102193 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.122829 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://569b1c6f63ead2d0fd069f4e3b144e45e7bb6f74d210c9a276df8d5d48cb51a7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:57Z\\\",\\\"message\\\":\\\"ector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.947520 6737 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:57.946905 6737 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0313 14:06:57.949275 6737 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:06:57.949375 6737 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:06:57.949412 6737 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 14:06:57.949440 6737 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0313 14:06:57.949483 6737 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0313 14:06:57.949481 6737 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:06:57.949508 6737 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0313 14:06:57.949515 6737 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:06:57.949523 6737 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:06:57.949543 6737 factory.go:656] Stopping watch factory\\\\nI0313 14:06:57.949567 6737 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.141229 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152262 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152347 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.152661 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bg45\" (UniqueName: \"kubernetes.io/projected/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-kube-api-access-6bg45\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.153178 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.153268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.160363 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.166225 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.170506 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bg45\" (UniqueName: \"kubernetes.io/projected/e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be-kube-api-access-6bg45\") pod \"ovnkube-control-plane-749d76644c-qncw4\" (UID: \"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175408 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175575 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175661 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175757 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.175860 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.180911 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.228352 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.243055 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.255977 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279026 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279104 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279146 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.279189 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.342783 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.371583 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385403 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385443 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385454 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385471 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.385483 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.386705 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.387099 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.405936 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.432685 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.447980 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.468541 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.486194 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.487953 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.487991 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.488006 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.488030 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.488051 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.501385 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.512980 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.529299 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.546540 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.561515 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.576539 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590564 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590609 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590621 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590642 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.590656 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.591073 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.604092 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.615469 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694385 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694504 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694517 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694535 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.694547 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783459 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783535 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783665 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.783731 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.783916 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-tzmvc"] Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.783953 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.784195 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.785232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.785329 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.796953 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.796996 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.797007 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.797025 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.797038 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.809863 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.831877 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.847386 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.860303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.860352 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdzl9\" (UniqueName: \"kubernetes.io/projected/f797b1bb-e769-49c4-84ce-93a6fad2df93-kube-api-access-tdzl9\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.860678 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.875781 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.897958 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900145 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900180 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900191 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900211 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.900228 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:00Z","lastTransitionTime":"2026-03-13T14:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.914277 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.930075 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.945061 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.961656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.961725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdzl9\" (UniqueName: \"kubernetes.io/projected/f797b1bb-e769-49c4-84ce-93a6fad2df93-kube-api-access-tdzl9\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.961933 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:00 crc kubenswrapper[4907]: E0313 14:07:00.962076 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:01.462036991 +0000 UTC m=+120.361824720 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.962673 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.981509 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:00Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:00 crc kubenswrapper[4907]: I0313 14:07:00.994977 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdzl9\" (UniqueName: \"kubernetes.io/projected/f797b1bb-e769-49c4-84ce-93a6fad2df93-kube-api-access-tdzl9\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003389 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003451 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003466 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003491 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003518 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.003619 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.030469 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.050895 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.077680 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106460 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106497 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106506 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106521 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.106531 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208612 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208662 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208675 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208696 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.208710 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311682 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311717 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311730 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311750 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.311769 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.391562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" event={"ID":"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be","Type":"ContainerStarted","Data":"540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.391624 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" event={"ID":"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be","Type":"ContainerStarted","Data":"07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.391636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" event={"ID":"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be","Type":"ContainerStarted","Data":"cdd74771efe08d942155d906ba03e386875af19b9e9f5488d227dd5f4c8b84ab"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415518 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415562 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415573 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415592 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.415603 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.417010 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.434744 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.448689 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.462132 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.466776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.467017 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.467131 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:02.467099639 +0000 UTC m=+121.366887348 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.477244 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.488979 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.500864 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.516118 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517827 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517853 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.517862 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.534640 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.549424 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.575028 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.592789 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.608329 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620406 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620528 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620677 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620756 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.620852 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:01Z","lastTransitionTime":"2026-03-13T14:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.621310 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.634664 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.721250 4907 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.805785 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.824253 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.838173 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.851769 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.867058 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: E0313 14:07:01.885603 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.887164 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.909593 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.927869 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.943949 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.957634 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.971769 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:01 crc kubenswrapper[4907]: I0313 14:07:01.989087 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.009639 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.025140 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.047479 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.479581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.479760 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.479868 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:04.479833493 +0000 UTC m=+123.379621212 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782103 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782211 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782248 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:02 crc kubenswrapper[4907]: I0313 14:07:02.782280 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782248 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782387 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782542 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:02 crc kubenswrapper[4907]: E0313 14:07:02.782822 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.503963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.504277 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.505325 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:08.505276072 +0000 UTC m=+127.405063971 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.605836 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.606193 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.606147345 +0000 UTC m=+155.505935084 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.706953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.707015 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.707041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.707060 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707148 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707202 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707185103 +0000 UTC m=+155.606972792 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707381 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707419 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707468 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707497 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707521 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707488232 +0000 UTC m=+155.607275951 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707594 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707563214 +0000 UTC m=+155.607350943 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707712 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707736 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707751 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.707810 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:36.707794622 +0000 UTC m=+155.607582351 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.782362 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.782426 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.782363 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.782530 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.782708 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.782945 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:04 crc kubenswrapper[4907]: I0313 14:07:04.783059 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:04 crc kubenswrapper[4907]: E0313 14:07:04.783291 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:05 crc kubenswrapper[4907]: I0313 14:07:05.164991 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.781625 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.781674 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.781716 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.781789 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.781970 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.782090 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:06 crc kubenswrapper[4907]: I0313 14:07:06.782364 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.782606 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:06 crc kubenswrapper[4907]: E0313 14:07:06.887840 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.018928 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019194 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019294 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019375 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.019450 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.034816 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039812 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039913 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039940 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039972 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.039996 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.061282 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066506 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066580 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066601 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066629 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.066647 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.087295 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092382 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092435 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092452 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092479 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.092497 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.110339 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115638 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115698 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115710 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115732 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:07 crc kubenswrapper[4907]: I0313 14:07:07.115770 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:07Z","lastTransitionTime":"2026-03-13T14:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.133013 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:07Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:07 crc kubenswrapper[4907]: E0313 14:07:07.133360 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.558695 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.559017 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.559169 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:16.55913248 +0000 UTC m=+135.458920209 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.781970 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.782047 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.781970 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:08 crc kubenswrapper[4907]: I0313 14:07:08.782024 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782210 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782344 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782451 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:08 crc kubenswrapper[4907]: E0313 14:07:08.782561 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.197344 4907 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781455 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.781570 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781627 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.781763 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.781834 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.782083 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:10 crc kubenswrapper[4907]: E0313 14:07:10.782660 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:10 crc kubenswrapper[4907]: I0313 14:07:10.792450 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.805134 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.824358 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.838833 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.855745 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.874521 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: E0313 14:07:11.888627 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.897750 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.922764 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.940808 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.962835 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:11 crc kubenswrapper[4907]: I0313 14:07:11.985052 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.005003 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.018181 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.032943 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.049438 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.064630 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.089048 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.781624 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.781697 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.781784 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.781844 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.782402 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.782726 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:12 crc kubenswrapper[4907]: I0313 14:07:12.783016 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:12 crc kubenswrapper[4907]: E0313 14:07:12.783281 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.782319 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.782470 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783052 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783113 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:14 crc kubenswrapper[4907]: I0313 14:07:14.783149 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.783177 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.783287 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:14 crc kubenswrapper[4907]: E0313 14:07:14.783381 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.170665 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.194157 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.206089 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.223596 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.236720 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.247749 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.256161 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.268100 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.281123 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.297620 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.324958 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.345021 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.358780 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.368894 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.378922 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.391263 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.406347 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.454374 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.456750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2"} Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.457166 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.472594 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.482571 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.492988 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.504909 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.513763 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.524850 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.535604 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.544981 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.554480 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.566033 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.579681 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.591803 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.605904 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.624545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.637722 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:15 crc kubenswrapper[4907]: I0313 14:07:15.646721 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:15Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.462228 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.462812 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/1.log" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.466845 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" exitCode=1 Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.466927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2"} Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.466975 4907 scope.go:117] "RemoveContainer" containerID="987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.467907 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.468120 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.488032 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.508064 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.519291 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.530704 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.544667 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.561774 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.577374 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.590428 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.603912 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.615850 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.627927 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.640014 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.652011 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.652448 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.652524 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.652625 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:07:32.652597329 +0000 UTC m=+151.552385058 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.666999 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.678865 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.700709 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://987fce3ea49cd6a0ce7b37820e7cf26ee21b0721b0b5ffda8e9e0f45bb0d22e3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:06:59Z\\\",\\\"message\\\":\\\"r *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.164308 6921 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0313 14:06:59.164611 6921 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.164770 6921 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 14:06:59.165234 6921 factory.go:656] Stopping watch factory\\\\nI0313 14:06:59.165255 6921 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 14:06:59.170923 6921 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0313 14:06:59.170953 6921 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0313 14:06:59.171016 6921 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:06:59.171061 6921 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:06:59.171151 6921 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:16Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782063 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782119 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782084 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:16 crc kubenswrapper[4907]: I0313 14:07:16.782071 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782203 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782309 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782407 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.782501 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:16 crc kubenswrapper[4907]: E0313 14:07:16.890072 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447184 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447227 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447252 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.447262 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.458388 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461820 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461913 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461937 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461960 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.461973 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.471041 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.474144 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.475420 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.475602 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477708 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477724 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477741 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.477755 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.488294 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.488917 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492060 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492092 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492104 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492121 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.492133 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.503593 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.510373 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514583 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514728 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514745 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514772 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.514792 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:17Z","lastTransitionTime":"2026-03-13T14:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.518824 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.525816 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: E0313 14:07:17.526089 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.528295 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.538843 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.550234 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.560061 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.571152 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.590845 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.600601 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.614594 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.629602 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.639552 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.653584 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.666800 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:17 crc kubenswrapper[4907]: I0313 14:07:17.676322 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:17Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.781662 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.781947 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.782062 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:18 crc kubenswrapper[4907]: I0313 14:07:18.782074 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782059 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782198 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782308 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:18 crc kubenswrapper[4907]: E0313 14:07:18.782380 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782493 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782537 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782539 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.782683 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:20 crc kubenswrapper[4907]: I0313 14:07:20.782712 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.782925 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.783116 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:20 crc kubenswrapper[4907]: E0313 14:07:20.783191 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.792335 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.801831 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.813891 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.824048 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.841851 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.856007 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.869678 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.881308 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: E0313 14:07:21.890707 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.896126 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.910661 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.924006 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.936392 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.951858 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.965671 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.981405 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:21 crc kubenswrapper[4907]: I0313 14:07:21.996426 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:21Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.781341 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.781480 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.781341 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.781360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.781841 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.781955 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:22 crc kubenswrapper[4907]: I0313 14:07:22.782062 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:22 crc kubenswrapper[4907]: E0313 14:07:22.782140 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781436 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781458 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:24 crc kubenswrapper[4907]: I0313 14:07:24.781396 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.781585 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.781725 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.781859 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:24 crc kubenswrapper[4907]: E0313 14:07:24.782028 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782222 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782266 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782420 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:26 crc kubenswrapper[4907]: I0313 14:07:26.782582 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782598 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782746 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.782826 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:26 crc kubenswrapper[4907]: E0313 14:07:26.891536 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.837877 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.838875 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.839290 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.839607 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.839932 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.860066 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865802 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865852 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865863 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865900 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.865913 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.890088 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895662 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895717 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895737 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895760 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.895778 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.918045 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924171 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924219 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924237 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924261 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.924277 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.944493 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949448 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949546 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949563 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949586 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:27 crc kubenswrapper[4907]: I0313 14:07:27.949603 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:27Z","lastTransitionTime":"2026-03-13T14:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.970052 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:27Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:27 crc kubenswrapper[4907]: E0313 14:07:27.970580 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782236 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782261 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.782389 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.782633 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.782817 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.783273 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.783332 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:28 crc kubenswrapper[4907]: I0313 14:07:28.783747 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:28 crc kubenswrapper[4907]: E0313 14:07:28.784156 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:07:29 crc kubenswrapper[4907]: I0313 14:07:29.798497 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781347 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781366 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:30 crc kubenswrapper[4907]: I0313 14:07:30.781971 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782270 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782612 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782784 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:30 crc kubenswrapper[4907]: E0313 14:07:30.782696 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.803310 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.823697 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.843604 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.862806 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.880822 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: E0313 14:07:31.892325 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.901126 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.920482 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.940170 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.954061 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.965824 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:31 crc kubenswrapper[4907]: I0313 14:07:31.987853 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:31Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.003051 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.022034 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.038022 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.053173 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.068456 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.084093 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:32Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.743875 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.744037 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.744088 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:04.744075243 +0000 UTC m=+183.643862932 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.781796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.781827 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.781911 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782005 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:32 crc kubenswrapper[4907]: I0313 14:07:32.782024 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782218 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:32 crc kubenswrapper[4907]: E0313 14:07:32.782302 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781387 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.781875 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781439 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.781985 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:34 crc kubenswrapper[4907]: I0313 14:07:34.781470 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.782082 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:34 crc kubenswrapper[4907]: E0313 14:07:34.782311 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.545937 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/0.log" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.546034 4907 generic.go:334] "Generic (PLEG): container finished" podID="64ebd338-0158-44dc-90ab-e7c285e87762" containerID="391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2" exitCode=1 Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.546101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerDied","Data":"391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2"} Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.547028 4907 scope.go:117] "RemoveContainer" containerID="391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.567665 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.587745 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.603142 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.618972 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.641951 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.658422 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.674439 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.690018 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.703952 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.718418 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.737440 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.758690 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.779046 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.795845 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.813251 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.826201 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:35 crc kubenswrapper[4907]: I0313 14:07:35.839401 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:35Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.557792 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/0.log" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.557910 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b"} Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.583674 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.606989 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.628910 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.648988 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.661920 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.675559 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.688955 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.697116 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.697661 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.697615688 +0000 UTC m=+219.597403377 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.709791 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.742488 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.764230 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781858 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.781916 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782082 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782277 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782425 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.782540 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.784210 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.798782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.798959 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.798999 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799016 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799029 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799042 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799087 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799057 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799049 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799026 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799005651 +0000 UTC m=+219.698793360 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799375 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799331735 +0000 UTC m=+219.699119474 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799418 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799403933 +0000 UTC m=+219.699191662 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.799465 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:08:40.799448567 +0000 UTC m=+219.699236296 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.811455 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.829040 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.848995 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.866643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.883793 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:36 crc kubenswrapper[4907]: E0313 14:07:36.894703 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:36 crc kubenswrapper[4907]: I0313 14:07:36.902248 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:36Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145725 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145775 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145793 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145818 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.145835 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.171276 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176929 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176948 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176976 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.176993 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.199763 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205427 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205481 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205500 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205524 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.205543 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.224732 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229348 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229402 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229416 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229439 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.229453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.245055 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249376 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249436 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249453 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249480 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.249498 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:38Z","lastTransitionTime":"2026-03-13T14:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.267639 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:38Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.267850 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782199 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782248 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782274 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.782412 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782516 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782667 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782774 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:38 crc kubenswrapper[4907]: E0313 14:07:38.782833 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:38 crc kubenswrapper[4907]: I0313 14:07:38.796505 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782275 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782384 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782227 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782512 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:40 crc kubenswrapper[4907]: I0313 14:07:40.782557 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782669 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:40 crc kubenswrapper[4907]: E0313 14:07:40.782741 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.782525 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.804866 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.824039 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.841087 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.856727 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.872448 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.886035 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: E0313 14:07:41.895398 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.903010 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.918636 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.935334 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.950925 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.968905 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:41 crc kubenswrapper[4907]: I0313 14:07:41.981549 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.001655 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:41Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.023817 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.039817 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.055220 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.065556 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.078097 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.582990 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.586419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.586864 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.604895 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.615338 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.625216 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.639838 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.649982 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.666139 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.682400 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.692204 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.701874 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.712116 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.722900 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.733855 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.745799 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.757538 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.769414 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782377 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782457 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782562 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.782617 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.782632 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.782694 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.782967 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:42 crc kubenswrapper[4907]: E0313 14:07:42.783129 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.795842 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:42 crc kubenswrapper[4907]: I0313 14:07:42.808404 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:42Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:43 crc kubenswrapper[4907]: I0313 14:07:43.795955 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782171 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782055 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782291 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782314 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:44 crc kubenswrapper[4907]: I0313 14:07:44.782336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:44 crc kubenswrapper[4907]: E0313 14:07:44.782399 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781800 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781871 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:46 crc kubenswrapper[4907]: I0313 14:07:46.781848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.781948 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.782026 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.782089 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.782134 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:46 crc kubenswrapper[4907]: E0313 14:07:46.989943 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.400792 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" probeResult="failure" output="" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543392 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543451 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543468 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543490 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.543506 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.556500 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559735 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559776 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559788 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559805 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.559817 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.577245 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581395 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581473 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581497 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581529 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.581554 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.596317 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600234 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600273 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600287 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600307 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.600320 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.618017 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621838 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621873 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621895 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621908 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.621919 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:48Z","lastTransitionTime":"2026-03-13T14:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.634359 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:48Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.634504 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781592 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781650 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781614 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.781741 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:48 crc kubenswrapper[4907]: I0313 14:07:48.781807 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.781959 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.781998 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:48 crc kubenswrapper[4907]: E0313 14:07:48.782067 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781499 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781545 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:50 crc kubenswrapper[4907]: I0313 14:07:50.781629 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781623 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781723 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781864 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:50 crc kubenswrapper[4907]: E0313 14:07:50.781922 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.803200 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.824563 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.842691 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.869066 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.921647 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.938813 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.953438 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.965784 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.975636 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: I0313 14:07:51.987553 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:51 crc kubenswrapper[4907]: E0313 14:07:51.990968 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.001266 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:51Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.017471 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.029071 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.055336 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.086465 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.108641 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.128425 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.144561 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.161034 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:52Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782062 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782144 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:52 crc kubenswrapper[4907]: I0313 14:07:52.782084 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.782372 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.782539 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.782873 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:52 crc kubenswrapper[4907]: E0313 14:07:52.783065 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.781608 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.781673 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.781835 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.781996 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:54 crc kubenswrapper[4907]: I0313 14:07:54.782042 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.782205 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.782305 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:54 crc kubenswrapper[4907]: E0313 14:07:54.782412 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782224 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782414 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782466 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782531 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:56 crc kubenswrapper[4907]: I0313 14:07:56.782571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.782649 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:56 crc kubenswrapper[4907]: E0313 14:07:56.992963 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.782418 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.782540 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.782670 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.782934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.783131 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.783299 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.783392 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.783455 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913548 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913632 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913658 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913693 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.913717 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.936196 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942223 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942281 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942300 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942328 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.942347 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.964148 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.969912 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.969970 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.969991 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.970020 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.970040 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:58 crc kubenswrapper[4907]: E0313 14:07:58.992728 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:58Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.998966 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999025 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999041 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999066 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:58 crc kubenswrapper[4907]: I0313 14:07:58.999087 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:58Z","lastTransitionTime":"2026-03-13T14:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:59 crc kubenswrapper[4907]: E0313 14:07:59.020542 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026297 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026356 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026372 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026399 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:07:59 crc kubenswrapper[4907]: I0313 14:07:59.026418 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:07:59Z","lastTransitionTime":"2026-03-13T14:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:07:59 crc kubenswrapper[4907]: E0313 14:07:59.048007 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:07:59Z is after 2025-08-24T17:21:41Z" Mar 13 14:07:59 crc kubenswrapper[4907]: E0313 14:07:59.048300 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781828 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:00 crc kubenswrapper[4907]: I0313 14:08:00.781996 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782034 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782097 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782265 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:00 crc kubenswrapper[4907]: E0313 14:08:00.782461 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.805158 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.819116 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.831327 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.843244 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.854430 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.868526 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.885937 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.905691 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.929268 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.942029 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.954961 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.970131 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: I0313 14:08:01.984226 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:01 crc kubenswrapper[4907]: E0313 14:08:01.994292 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.002044 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:01Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.036083 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.054124 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.075208 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.091690 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.111107 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:02Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.781505 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.781706 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.782066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.782132 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:02 crc kubenswrapper[4907]: I0313 14:08:02.782172 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.782319 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.782495 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:02 crc kubenswrapper[4907]: E0313 14:08:02.782593 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.673350 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.674474 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/2.log" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.687504 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" exitCode=1 Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.687564 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.687615 4907 scope.go:117] "RemoveContainer" containerID="a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.688801 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:03 crc kubenswrapper[4907]: E0313 14:08:03.689216 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.729115 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:08:02Z\\\",\\\"message\\\":\\\"dn-host-isolation-manager: error found while processing openshift-dns-operator/dns-operator-744455d44c-xkdjh: failed to check if pod openshift-dns-operator/dns-operator-744455d44c-xkdjh is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0313 14:08:02.544936 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI0313 14:08:02.544957 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-multus/network-metrics-daemon-tzmvc: failed to check if pod openshift-multus/network-metrics-daemon-tzmvc is in primary UDN: could not find OVN pod annotation in map[cluster-autoscaler.kubernetes.io/enable-ds-eviction:false]\\\\nE0313 14:08:02.629503 7501 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI0313 14:08:02.630816 7501 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:08:02.630949 7501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.747752 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.764313 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.785779 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.812618 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.831331 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.863534 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.882037 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.901813 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.916073 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.929797 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.946141 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.963022 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:03 crc kubenswrapper[4907]: I0313 14:08:03.985403 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.002443 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:03Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.021655 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.035233 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.051281 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.069579 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:04Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.694197 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.781962 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.782007 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.782047 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.782067 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782155 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782263 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782469 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.782528 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:04 crc kubenswrapper[4907]: I0313 14:08:04.835674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.835918 4907 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:08:04 crc kubenswrapper[4907]: E0313 14:08:04.836061 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs podName:f797b1bb-e769-49c4-84ce-93a6fad2df93 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.836021758 +0000 UTC m=+247.735809497 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs") pod "network-metrics-daemon-tzmvc" (UID: "f797b1bb-e769-49c4-84ce-93a6fad2df93") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781811 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781811 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782072 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782165 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:06 crc kubenswrapper[4907]: I0313 14:08:06.781844 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782272 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.782508 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:06 crc kubenswrapper[4907]: E0313 14:08:06.995989 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.782057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.782073 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.782232 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.782337 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.782079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.782522 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:08 crc kubenswrapper[4907]: I0313 14:08:08.783137 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:08 crc kubenswrapper[4907]: E0313 14:08:08.783299 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298160 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298172 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298196 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.298209 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.319100 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.325982 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326049 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326069 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326097 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.326117 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.350203 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355576 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355655 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355675 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355707 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.355727 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.375858 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.381674 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.381935 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.382109 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.382280 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.382453 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.406198 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412399 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412503 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412523 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412558 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:09 crc kubenswrapper[4907]: I0313 14:08:09.412579 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:09Z","lastTransitionTime":"2026-03-13T14:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.435342 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T14:08:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4b182bae-7f2d-4b54-934c-85a9a277d132\\\",\\\"systemUUID\\\":\\\"05bfbfa3-f04f-45ba-8233-a355fc774c24\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:09Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:09 crc kubenswrapper[4907]: E0313 14:08:09.435570 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781550 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781629 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:10 crc kubenswrapper[4907]: I0313 14:08:10.781731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.781986 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.782353 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.782441 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:10 crc kubenswrapper[4907]: E0313 14:08:10.782611 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.803119 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4819028f34db81cc43a04e4854c30488ed876c63b8daff11105bf61546c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.822643 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.843527 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xtwf7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"64ebd338-0158-44dc-90ab-e7c285e87762\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:34Z\\\",\\\"message\\\":\\\"2026-03-13T14:06:49+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531\\\\n2026-03-13T14:06:49+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_baa16ed5-a467-4fd5-8846-0c869be5e531 to /host/opt/cni/bin/\\\\n2026-03-13T14:06:49Z [verbose] multus-daemon started\\\\n2026-03-13T14:06:49Z [verbose] Readiness Indicator file check\\\\n2026-03-13T14:07:34Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6cwbq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xtwf7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.864495 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bed3bcea-e8c0-4b00-a985-e0c5b5bea100\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://402c1ef89085d840edb4313577f5b2fb967072cb3db00c34c98572f00699529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22becea3c68103ff0b2639c7ae8c051b76d19b407b0463645fc2e4a0682f23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:05:32Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 14:05:03.905343 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 14:05:03.908030 1 observer_polling.go:159] Starting file observer\\\\nI0313 14:05:03.938230 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 14:05:03.942399 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 14:05:32.345768 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 14:05:32.345932 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:05:31Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abfa5a9ad1a98a2907ead28390ee926e176646fe761b89f580101bd429cf747\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc67477e0223db1e4b403be3709fe71fe6409570361e103eb730fdeaffe3e80\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.882914 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.900545 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afb1d430459aafab2d6bc2507b4b4059529401e51fdf8e9848b0d7bb8651db68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.918515 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-745zx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58e11265-2fd9-4b9b-926c-3343bb84ee53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ce7ad32fd2bf2332f3ee41080115b083c34326b7dc31c950098dfba15f58422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mnz7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-745zx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.935581 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f797b1bb-e769-49c4-84ce-93a6fad2df93\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdzl9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-tzmvc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.960082 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1006376d-4033-454e-8a23-2b67882bdd0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a99cabe30a7abc8da854bf86d79628b94a6ffc20c28f828b695080332b319db2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"message\\\":\\\"Sending *v1.Node event handler 7 for removal\\\\nI0313 14:07:15.619307 7162 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0313 14:07:15.619320 7162 handler.go:208] Removed *v1.Node event handler 2\\\\nI0313 14:07:15.619329 7162 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 14:07:15.619343 7162 handler.go:208] Removed *v1.Node event handler 7\\\\nI0313 14:07:15.619351 7162 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 14:07:15.619358 7162 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 14:07:15.619362 7162 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0313 14:07:15.619381 7162 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 14:07:15.619382 7162 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 14:07:15.619392 7162 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 14:07:15.619619 7162 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0313 14:07:15.619713 7162 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0313 14:07:15.619975 7162 factory.go:656] Stopping watch factory\\\\nI0313 14:07:15.620001 7162 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:07:15.620033 7162 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 14:07:15.620096 7162 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T14:08:02Z\\\",\\\"message\\\":\\\"dn-host-isolation-manager: error found while processing openshift-dns-operator/dns-operator-744455d44c-xkdjh: failed to check if pod openshift-dns-operator/dns-operator-744455d44c-xkdjh is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0313 14:08:02.544936 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx: failed to check if pod openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:nonroot-v2 openshift.io/scc:nonroot-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nI0313 14:08:02.544957 7501 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-multus/network-metrics-daemon-tzmvc: failed to check if pod openshift-multus/network-metrics-daemon-tzmvc is in primary UDN: could not find OVN pod annotation in map[cluster-autoscaler.kubernetes.io/enable-ds-eviction:false]\\\\nE0313 14:08:02.629503 7501 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI0313 14:08:02.630816 7501 ovnkube.go:599] Stopped ovnkube\\\\nI0313 14:08:02.630949 7501 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:07:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tm9lw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qxlx5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.976027 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a21d7e2-0481-4ea4-b129-0e2754315289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://35fd0db86ae409203bc57f755e489cd542ea68d19b5fcb9023794a77bcfa1a64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22af7218ec732d25874992d8015afd1d73b48cb6cf7d74e638c9db396cb9efe3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: I0313 14:08:11.987954 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584d221c-be74-4a39-8173-6c8e3f55f686\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b782a795257057929fe616938d1483a9accbd7a93f0cba963f9791a3d4ec4749\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c826e7b9be5e3a0ecdaa6dfeaf2454535e542ff86acc532cc3c80acf4a95c1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4333bdf047392cb3f64e29e677fe8d2f21c7ebca1a7e7097167e70c6785b8d46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a4dea84b35a82ce16f0e0d38057dc994b26240cf45dcc9111ed0de573a62881\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:11 crc kubenswrapper[4907]: E0313 14:08:11.997459 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.000062 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:32Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:11Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.016582 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"57e4e912-9a98-4419-87ce-f16e178ed921\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38f113418d62dab1728f0ba57cd1d289e3ec41621a338017e74c9996a8449b1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05bf7ccdc21826b7565ea477ae9c624002c5f9640d7ebec506ba3fae3605c10f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e70e926373e1ed98bc7c03c9ff58c84704b568ddc5da120f845085e48974b302\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f083ca42d37d6d3eecb9d55cf639617fa79f6a89d81286053d1a2a89c3346d45\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bdef68f401b5ba2c7cf61551eba416e74f295ab2281433a5115334d28094818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f8d1f40744d609fd964572ef5e323f0283d2582e98745602754165818f17847\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c1ab2e148eaa177b98f5883a3f323d12f5b662bff8cb2d072a97be18ab8e120\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:06:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vldr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-p8ptc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.027402 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f3fb4213-144f-4b01-9d78-ff3315761bb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba1140a2462713c2401b47a4cf1f231bc9d9ae7867a7d0b1b597214a1f7efe07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxwbn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hm56j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.050426 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"31a594bd-13e7-4c68-ac7c-f481626088ec\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8141bbe3b2b50ad7d2aed26f813a951bc6db4c90bf8271cd23a125cc7ec2eb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb841f47d97b3e9495658ff7f6aa3f835905a9cd2ee9736774c3504d6c59a209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f04995f26a4c320fa30f2a97b1299284cbb2e23965bae16a6cca8d06eef630eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0568aef54a902fedcaaa71cabe0add0cbeb3a38743a7fe45584c5e315292afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://171ea306a9ec518a88ef509adaab1a7c238abf7b5824a87d3425de7199cde520\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14a9ee19b96b2019de6ad9e5092e0004a4587a9f678fec417c4ac02e60695ac8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3d46350802ae58ed4c464dc3f8d3f9dd7e091ef8091659e0711b9d2632e00708\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65bd2db6ca4ffb26d9cb4bce44dc27a862353533bb0ec584bcc688b73fbec3eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.064794 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5d2b054-0ed8-419c-b731-f14d3598372f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:05:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T14:06:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 14:06:13.202871 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 14:06:13.203175 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 14:06:13.204427 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2471728650/tls.crt::/tmp/serving-cert-2471728650/tls.key\\\\\\\"\\\\nI0313 14:06:13.581160 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 14:06:13.584012 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 14:06:13.584036 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 14:06:13.584064 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 14:06:13.584072 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 14:06:13.594504 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0313 14:06:13.594891 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0313 14:06:13.595323 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595344 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 14:06:13.595356 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 14:06:13.595363 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 14:06:13.595370 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 14:06:13.595376 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0313 14:06:13.597226 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T14:06:12Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:05:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T14:05:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T14:05:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:05:01Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.078065 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94b045cf2fed56457a2ca14f489bb69470198f159f16c207fe061d034dba1555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd28f7a5c35822582b70ed41645294aaede10e28ff59f3f3b5d36e7da8237b76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.089106 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-l7zg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d105034b-f041-4fb5-8df6-6b1bb7681fda\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:06:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a207f917836785e320f4f79aca5d6449da2fdeec3ee0765f34d227d5617fa1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szmh9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:06:47Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-l7zg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.103310 4907 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1d7ee39-09d7-4c0e-942b-5e9d8c9f63be\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T14:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07df71887196cc7098ac9606a4a15b89d45219048c31280198ae75022bff9370\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://540ca207f31c9a8a9892f80dee153b3fb86f8de406f38360221512f424d8c884\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T14:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6bg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T14:07:00Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qncw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T14:08:12Z is after 2025-08-24T17:21:41Z" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.781860 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.781942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.782069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:12 crc kubenswrapper[4907]: I0313 14:08:12.782070 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.782256 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.782727 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.782859 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:12 crc kubenswrapper[4907]: E0313 14:08:12.783066 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782214 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782247 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782443 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:14 crc kubenswrapper[4907]: I0313 14:08:14.782479 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783171 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783207 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783359 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:14 crc kubenswrapper[4907]: E0313 14:08:14.783445 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.798789 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799002 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.799283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799373 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.799434 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799625 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:16 crc kubenswrapper[4907]: I0313 14:08:16.799687 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.799806 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:16 crc kubenswrapper[4907]: E0313 14:08:16.998605 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.781814 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.781847 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.781939 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.781974 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.782022 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.782125 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.782302 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.782686 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.783108 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:18 crc kubenswrapper[4907]: E0313 14:08:18.783317 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.816431 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=49.816414781 podStartE2EDuration="49.816414781s" podCreationTimestamp="2026-03-13 14:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.816289218 +0000 UTC m=+197.716076977" watchObservedRunningTime="2026-03-13 14:08:18.816414781 +0000 UTC m=+197.716202490" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.872142 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-745zx" podStartSLOduration=138.872107526 podStartE2EDuration="2m18.872107526s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.871831198 +0000 UTC m=+197.771618907" watchObservedRunningTime="2026-03-13 14:08:18.872107526 +0000 UTC m=+197.771895255" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.900807 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=35.900781476 podStartE2EDuration="35.900781476s" podCreationTimestamp="2026-03-13 14:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.899801189 +0000 UTC m=+197.799588918" watchObservedRunningTime="2026-03-13 14:08:18.900781476 +0000 UTC m=+197.800569175" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.917288 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=68.917269243 podStartE2EDuration="1m8.917269243s" podCreationTimestamp="2026-03-13 14:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.912419582 +0000 UTC m=+197.812207301" watchObservedRunningTime="2026-03-13 14:08:18.917269243 +0000 UTC m=+197.817056932" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.949958 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-p8ptc" podStartSLOduration=137.949931102 podStartE2EDuration="2m17.949931102s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.949808569 +0000 UTC m=+197.849596278" watchObservedRunningTime="2026-03-13 14:08:18.949931102 +0000 UTC m=+197.849718831" Mar 13 14:08:18 crc kubenswrapper[4907]: I0313 14:08:18.962758 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podStartSLOduration=137.96273609 podStartE2EDuration="2m17.96273609s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:18.962645018 +0000 UTC m=+197.862432727" watchObservedRunningTime="2026-03-13 14:08:18.96273609 +0000 UTC m=+197.862523779" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.019283 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=41.019261126 podStartE2EDuration="41.019261126s" podCreationTimestamp="2026-03-13 14:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.016848 +0000 UTC m=+197.916635709" watchObservedRunningTime="2026-03-13 14:08:19.019261126 +0000 UTC m=+197.919048835" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.038191 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=99.038163941 podStartE2EDuration="1m39.038163941s" podCreationTimestamp="2026-03-13 14:06:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.037753589 +0000 UTC m=+197.937541278" watchObservedRunningTime="2026-03-13 14:08:19.038163941 +0000 UTC m=+197.937951650" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.070574 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-l7zg7" podStartSLOduration=139.070554201 podStartE2EDuration="2m19.070554201s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.070445738 +0000 UTC m=+197.970233447" watchObservedRunningTime="2026-03-13 14:08:19.070554201 +0000 UTC m=+197.970341880" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.149243 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qncw4" podStartSLOduration=138.14922719 podStartE2EDuration="2m18.14922719s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.086528425 +0000 UTC m=+197.986316124" watchObservedRunningTime="2026-03-13 14:08:19.14922719 +0000 UTC m=+198.049014879" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.192027 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xtwf7" podStartSLOduration=138.192000823 podStartE2EDuration="2m18.192000823s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:19.1907857 +0000 UTC m=+198.090573409" watchObservedRunningTime="2026-03-13 14:08:19.192000823 +0000 UTC m=+198.091788542" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585135 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585195 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585211 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585235 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.585251 4907 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T14:08:19Z","lastTransitionTime":"2026-03-13T14:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.650939 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw"] Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.652078 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.655643 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.656053 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.655720 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.655844 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696005 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696121 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696236 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696278 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.696326 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.796940 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797111 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.797147 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.798057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.798097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.798785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.805746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.819517 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-h7pgw\" (UID: \"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.871477 4907 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.882676 4907 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 14:08:19 crc kubenswrapper[4907]: I0313 14:08:19.977038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.767155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" event={"ID":"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b","Type":"ContainerStarted","Data":"4ed47fad7c3de74b8b9b0a421e06356ab4fa2cf4bcb3e89c6c7d4963e5c44def"} Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.767658 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" event={"ID":"2d0f4392-a2c3-4ca7-9b92-b7ea2c01579b","Type":"ContainerStarted","Data":"dad1aa0752bc06328372e05af2c284010393fa1cd05fd1b800414d163c3d36b2"} Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781419 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781455 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.781917 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781553 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:20 crc kubenswrapper[4907]: I0313 14:08:20.781497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.782308 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.782026 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:20 crc kubenswrapper[4907]: E0313 14:08:20.782523 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773112 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773595 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/0.log" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773623 4907 generic.go:334] "Generic (PLEG): container finished" podID="64ebd338-0158-44dc-90ab-e7c285e87762" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" exitCode=1 Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerDied","Data":"ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b"} Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.773687 4907 scope.go:117] "RemoveContainer" containerID="391b8bc1e975982b172175fcf4b4509756b018cbb99c12e48f4931ff472cdae2" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.774086 4907 scope.go:117] "RemoveContainer" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" Mar 13 14:08:21 crc kubenswrapper[4907]: E0313 14:08:21.774215 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xtwf7_openshift-multus(64ebd338-0158-44dc-90ab-e7c285e87762)\"" pod="openshift-multus/multus-xtwf7" podUID="64ebd338-0158-44dc-90ab-e7c285e87762" Mar 13 14:08:21 crc kubenswrapper[4907]: I0313 14:08:21.811260 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-h7pgw" podStartSLOduration=141.811239521 podStartE2EDuration="2m21.811239521s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:20.78679233 +0000 UTC m=+199.686580029" watchObservedRunningTime="2026-03-13 14:08:21.811239521 +0000 UTC m=+200.711027220" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.000610 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.780946 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782113 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782193 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:22 crc kubenswrapper[4907]: I0313 14:08:22.782327 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.782514 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.782697 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.782854 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:22 crc kubenswrapper[4907]: E0313 14:08:22.783318 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782242 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782392 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782451 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782570 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:24 crc kubenswrapper[4907]: I0313 14:08:24.782601 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782657 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:24 crc kubenswrapper[4907]: E0313 14:08:24.782716 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.781744 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783303 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.782435 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.782384 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783501 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:26 crc kubenswrapper[4907]: I0313 14:08:26.782450 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783801 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:26 crc kubenswrapper[4907]: E0313 14:08:26.783866 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:27 crc kubenswrapper[4907]: E0313 14:08:27.002723 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.782141 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.782500 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.783055 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.783183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.783212 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:28 crc kubenswrapper[4907]: I0313 14:08:28.783383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.783432 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:28 crc kubenswrapper[4907]: E0313 14:08:28.783364 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782202 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782287 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782202 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782378 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782467 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:30 crc kubenswrapper[4907]: I0313 14:08:30.782294 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782587 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:30 crc kubenswrapper[4907]: E0313 14:08:30.782675 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:31 crc kubenswrapper[4907]: I0313 14:08:31.785094 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:31 crc kubenswrapper[4907]: E0313 14:08:31.785435 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qxlx5_openshift-ovn-kubernetes(1006376d-4033-454e-8a23-2b67882bdd0b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.004246 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.781932 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.782091 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.782040 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:32 crc kubenswrapper[4907]: I0313 14:08:32.782014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782274 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782453 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782604 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:32 crc kubenswrapper[4907]: E0313 14:08:32.782706 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:33 crc kubenswrapper[4907]: I0313 14:08:33.783733 4907 scope.go:117] "RemoveContainer" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782387 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782388 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.782874 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.782419 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.783216 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.783358 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:34 crc kubenswrapper[4907]: E0313 14:08:34.783522 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.831774 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:08:34 crc kubenswrapper[4907]: I0313 14:08:34.831865 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173"} Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782359 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782457 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:36 crc kubenswrapper[4907]: I0313 14:08:36.782408 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.782637 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.782820 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.783132 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:36 crc kubenswrapper[4907]: E0313 14:08:36.783397 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:37 crc kubenswrapper[4907]: E0313 14:08:37.005297 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781443 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.781662 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:38 crc kubenswrapper[4907]: I0313 14:08:38.781463 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.782032 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.782142 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:38 crc kubenswrapper[4907]: E0313 14:08:38.782315 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.768335 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.768651 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.768599129 +0000 UTC m=+341.668386858 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782204 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782273 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782377 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.782229 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782420 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782592 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782756 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.782987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870472 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870553 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870615 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:40 crc kubenswrapper[4907]: I0313 14:08:40.870674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870729 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870799 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870856 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870932 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870944 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.870831048 +0000 UTC m=+341.770618777 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870973 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.870993 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.870973872 +0000 UTC m=+341.770761601 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871048 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.871028564 +0000 UTC m=+341.770816283 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871171 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871244 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871273 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:40 crc kubenswrapper[4907]: E0313 14:08:40.871403 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:10:42.871367573 +0000 UTC m=+341.771155302 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.007471 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781385 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781436 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781436 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:42 crc kubenswrapper[4907]: I0313 14:08:42.781371 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.781529 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.781757 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.782015 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:42 crc kubenswrapper[4907]: E0313 14:08:42.782314 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782376 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782540 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784082 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:44 crc kubenswrapper[4907]: I0313 14:08:44.782589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784194 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784467 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:44 crc kubenswrapper[4907]: E0313 14:08:44.784671 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:45 crc kubenswrapper[4907]: I0313 14:08:45.783190 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.144936 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.150805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerStarted","Data":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.151592 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.191835 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podStartSLOduration=165.191798508 podStartE2EDuration="2m45.191798508s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:08:46.188790897 +0000 UTC m=+225.088578586" watchObservedRunningTime="2026-03-13 14:08:46.191798508 +0000 UTC m=+225.091586237" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781428 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.781537 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781563 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.781342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.781674 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.781782 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:46 crc kubenswrapper[4907]: E0313 14:08:46.782046 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:46 crc kubenswrapper[4907]: I0313 14:08:46.944104 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tzmvc"] Mar 13 14:08:47 crc kubenswrapper[4907]: E0313 14:08:47.010207 4907 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:08:47 crc kubenswrapper[4907]: I0313 14:08:47.154751 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:47 crc kubenswrapper[4907]: E0313 14:08:47.154988 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.041565 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.041676 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.782934 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782302 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.783078 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782401 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.783178 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:48 crc kubenswrapper[4907]: I0313 14:08:48.782241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:48 crc kubenswrapper[4907]: E0313 14:08:48.783264 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781519 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781608 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781664 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:50 crc kubenswrapper[4907]: I0313 14:08:50.781596 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.781773 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-tzmvc" podUID="f797b1bb-e769-49c4-84ce-93a6fad2df93" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.781991 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.782108 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:08:50 crc kubenswrapper[4907]: E0313 14:08:50.782260 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782169 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782131 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.782432 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.787803 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788170 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788321 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788379 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788485 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 14:08:52 crc kubenswrapper[4907]: I0313 14:08:52.788568 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.464995 4907 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.512943 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.513688 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.514634 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rnh5t"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.515178 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.515586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.515701 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.522547 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.523431 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.525010 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.525803 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.526381 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.526612 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.526997 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.527222 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.527426 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.527691 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528027 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528239 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528443 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528655 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.528952 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.529219 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.529469 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.529676 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-r86r4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.530645 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.531170 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-prlhd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.532351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.534249 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.535201 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550016 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550604 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550990 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.551160 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550709 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.551681 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.550784 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552070 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552208 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552284 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.552677 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553052 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553254 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553481 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553820 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553271 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.554211 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.569423 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.571598 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.574512 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553392 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553465 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553572 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.553693 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.554009 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.593083 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.593866 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594257 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594267 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594333 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594437 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594450 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.594612 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.596197 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597099 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-trusted-ca-bundle\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597160 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597197 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597224 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-policies\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597263 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-image-import-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597298 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005dfe54-e402-4aa4-8b67-f7e7b685debd-machine-approver-tls\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597364 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-service-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597386 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-serving-cert\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597433 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-config\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597456 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597507 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597580 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-client\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-dir\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597672 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jwmk\" (UniqueName: \"kubernetes.io/projected/b75dbfe3-8887-4a0b-9541-a4e4000924cb-kube-api-access-7jwmk\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597725 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597750 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597786 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597815 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-auth-proxy-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597855 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-client\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597876 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-encryption-config\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597949 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.597990 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598014 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3baeb4f-988d-4d42-b469-a310ca127745-serving-cert\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598034 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit-dir\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598078 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-289bg\" (UniqueName: \"kubernetes.io/projected/a3baeb4f-988d-4d42-b469-a310ca127745-kube-api-access-289bg\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598099 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598136 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598173 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598270 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.599073 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.599185 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.599765 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.600015 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.600341 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.600414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.601531 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.601947 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.602042 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-544df"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.602438 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.604101 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.604113 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.604545 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.598159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605643 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605680 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605704 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605739 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-serving-cert\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605763 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605785 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-encryption-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605811 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-config\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-serving-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605928 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b75dbfe3-8887-4a0b-9541-a4e4000924cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.605993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606012 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-images\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606033 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606078 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606243 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606296 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-node-pullsecrets\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.606696 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.609628 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w76kw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.610595 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.609089 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.611132 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xkdjh"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.611742 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.612163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.612503 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.612505 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.615950 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.617788 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.620098 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hwf6r"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.622542 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.621566 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.621737 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.621801 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623656 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623861 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623749 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.623822 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.624327 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-f8mmz"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.624825 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rnh5t"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.624998 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625132 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625264 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625381 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.625539 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.626106 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.626595 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.628390 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.629282 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.629387 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.632053 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.633309 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634317 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634554 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.634689 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.642732 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.643248 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.643658 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.646054 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.646820 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.647487 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.647628 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.648027 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.651232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.651678 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.651949 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.652664 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.652965 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653209 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653352 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653634 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.653908 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.654377 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687327 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687492 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687653 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.687922 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.689561 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.689997 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690416 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690455 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690772 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690984 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.691213 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.691739 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.692608 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.695271 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.696115 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.696815 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ksfb8"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.697379 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.697604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.698300 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.690874 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.705138 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706118 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706843 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-auth-proxy-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706877 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706917 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706944 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706967 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d44a1e-0eab-4bd6-833f-143436e43758-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.706987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707013 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-client\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707032 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707078 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3baeb4f-988d-4d42-b469-a310ca127745-serving-cert\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707129 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit-dir\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-encryption-config\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707217 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-289bg\" (UniqueName: \"kubernetes.io/projected/a3baeb4f-988d-4d42-b469-a310ca127745-kube-api-access-289bg\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707255 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707281 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707374 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707399 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707423 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae320273-06e8-43c5-a64f-acd80ad16d1c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707473 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-serving-cert\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707501 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707573 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-proxy-tls\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b66665-221d-4334-8263-aeb3b20e1c92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707619 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnmbh\" (UniqueName: \"kubernetes.io/projected/a7b66665-221d-4334-8263-aeb3b20e1c92-kube-api-access-rnmbh\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h5xn\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-kube-api-access-7h5xn\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707699 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707724 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-serving-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-encryption-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707770 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-config\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707786 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708310 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708627 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b75dbfe3-8887-4a0b-9541-a4e4000924cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae320273-06e8-43c5-a64f-acd80ad16d1c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708679 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-auth-proxy-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-images\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708727 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708775 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708810 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708930 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708949 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b66665-221d-4334-8263-aeb3b20e1c92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.708967 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-node-pullsecrets\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.709058 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.709842 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.709865 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711415 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711858 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.711961 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.712390 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.719819 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.720495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005dfe54-e402-4aa4-8b67-f7e7b685debd-config\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.721260 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.721362 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-serving-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.721852 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.730683 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.731967 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-config\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.731875 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a3baeb4f-988d-4d42-b469-a310ca127745-serving-cert\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.732214 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-encryption-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.732436 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-client\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.732494 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-node-pullsecrets\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.734483 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-config\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707446 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.735132 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b75dbfe3-8887-4a0b-9541-a4e4000924cb-images\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.736460 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.737104 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.707515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit-dir\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.737944 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.738278 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.738422 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-trusted-ca-bundle\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.739014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740002 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-trusted-ca-bundle\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740122 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-policies\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740170 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/963f6709-e807-40bb-aa9e-9bba24e82b71-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-policies\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740701 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-image-import-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.740727 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dslzm\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-kube-api-access-dslzm\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741320 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d44a1e-0eab-4bd6-833f-143436e43758-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741665 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44a1e-0eab-4bd6-833f-143436e43758-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.741736 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005dfe54-e402-4aa4-8b67-f7e7b685debd-machine-approver-tls\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-service-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-serving-cert\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742419 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742722 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-service-ca-bundle\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.742839 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.743227 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.743605 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/b75dbfe3-8887-4a0b-9541-a4e4000924cb-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.743868 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744356 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744409 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-image-import-ca\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744539 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.744786 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745071 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-config\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.745700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a9c889b-5954-4d3a-ae72-12000ee1cc25-audit\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746616 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746672 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-client\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-dir\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.746816 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747054 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747119 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-audit-dir\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jwmk\" (UniqueName: \"kubernetes.io/projected/b75dbfe3-8887-4a0b-9541-a4e4000924cb-kube-api-access-7jwmk\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747220 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747262 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb574\" (UniqueName: \"kubernetes.io/projected/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-kube-api-access-jb574\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747357 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747413 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.747439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/963f6709-e807-40bb-aa9e-9bba24e82b71-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.749266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.750494 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.758488 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-75gdt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760089 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760114 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760579 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.760822 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.761909 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a3baeb4f-988d-4d42-b469-a310ca127745-config\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.761941 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.776996 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-etcd-client\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.777134 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.777502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/005dfe54-e402-4aa4-8b67-f7e7b685debd-machine-approver-tls\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.778224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-encryption-config\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.778837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779108 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779469 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a9c889b-5954-4d3a-ae72-12000ee1cc25-serving-cert\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779617 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.779939 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.780266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.780665 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-serving-cert\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.781110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.781241 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.781355 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.787477 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.788653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.789248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.790621 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.790819 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zwzc4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.791770 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.794714 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.796440 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.797672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.799528 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.803452 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w76kw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.804817 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.805893 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.810698 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.810738 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.813220 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-g6q2r"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.814511 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.814540 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.814687 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.815662 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-544df"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.816367 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.822567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.822630 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.824674 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-r86r4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.827236 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.829025 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-prlhd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.830986 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.832469 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.835068 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.835222 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xkdjh"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.836808 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hwf6r"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.837993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.838951 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-252nq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.839791 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.840027 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sjwvn"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.841147 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.841276 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.842320 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ksfb8"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.844457 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.846798 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-75gdt"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848112 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b66665-221d-4334-8263-aeb3b20e1c92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/963f6709-e807-40bb-aa9e-9bba24e82b71-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dslzm\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-kube-api-access-dslzm\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d44a1e-0eab-4bd6-833f-143436e43758-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44a1e-0eab-4bd6-833f-143436e43758-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848255 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb574\" (UniqueName: \"kubernetes.io/projected/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-kube-api-access-jb574\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848278 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/963f6709-e807-40bb-aa9e-9bba24e82b71-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848315 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848330 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d44a1e-0eab-4bd6-833f-143436e43758-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848366 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848382 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848434 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae320273-06e8-43c5-a64f-acd80ad16d1c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-proxy-tls\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848469 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b66665-221d-4334-8263-aeb3b20e1c92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848485 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnmbh\" (UniqueName: \"kubernetes.io/projected/a7b66665-221d-4334-8263-aeb3b20e1c92-kube-api-access-rnmbh\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h5xn\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-kube-api-access-7h5xn\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848525 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848563 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae320273-06e8-43c5-a64f-acd80ad16d1c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.848597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.849066 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.849463 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-8jqgq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.849647 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b66665-221d-4334-8263-aeb3b20e1c92-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.850283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.850465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.850792 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/963f6709-e807-40bb-aa9e-9bba24e82b71-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.851582 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.853070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.853116 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.853512 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/963f6709-e807-40bb-aa9e-9bba24e82b71-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.854245 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b66665-221d-4334-8263-aeb3b20e1c92-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.854519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.854668 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-252nq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.855150 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.856159 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.857115 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.857701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.858330 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.860128 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.861331 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.862834 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.864697 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8jqgq"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.865975 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sjwvn"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.867261 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zwzc4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.868526 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.875429 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.895915 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.915543 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.935423 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.955641 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.975549 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 14:09:00 crc kubenswrapper[4907]: I0313 14:09:00.996666 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.016857 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.035645 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.056468 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.076181 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.096596 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.115944 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.135828 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.143074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ae320273-06e8-43c5-a64f-acd80ad16d1c-metrics-tls\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.160792 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.172818 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae320273-06e8-43c5-a64f-acd80ad16d1c-trusted-ca\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.175533 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.215926 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.236151 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.256841 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.276221 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.296707 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.316145 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.336733 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.355564 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.377095 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.385435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-proxy-tls\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.396311 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.436708 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.456110 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.476261 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.497051 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.516031 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.536141 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.556680 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.576914 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.583219 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d44a1e-0eab-4bd6-833f-143436e43758-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.596819 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.600109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44a1e-0eab-4bd6-833f-143436e43758-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.616375 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.636366 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.657948 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.698372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.703525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-289bg\" (UniqueName: \"kubernetes.io/projected/a3baeb4f-988d-4d42-b469-a310ca127745-kube-api-access-289bg\") pod \"authentication-operator-69f744f599-rnh5t\" (UID: \"a3baeb4f-988d-4d42-b469-a310ca127745\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.714623 4907 request.go:700] Waited for 1.005964719s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.717587 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.796719 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.816476 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.836471 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.862865 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.876670 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.975368 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 14:09:01 crc kubenswrapper[4907]: I0313 14:09:01.996449 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.016553 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.035583 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.056989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.076380 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.096300 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.116468 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.156329 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.176566 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.196280 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.216117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.235066 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.255923 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.276100 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.295417 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.316322 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.335970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.356345 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.375269 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.395774 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.416266 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.435619 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.456275 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.476434 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.496248 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.516810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.536433 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.556975 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.577173 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.595836 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.617049 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.635453 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.656454 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.676519 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.696515 4907 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.715651 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.734432 4907 request.go:700] Waited for 1.88512071s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.751076 4907 projected.go:288] Couldn't get configMap openshift-oauth-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.757975 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.766990 4907 projected.go:288] Couldn't get configMap openshift-authentication/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.779571 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.789683 4907 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.811739 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.811838 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.829974 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb574\" (UniqueName: \"kubernetes.io/projected/a8e9ea4a-43ad-4da0-99fa-c01eea712c8c-kube-api-access-jb574\") pod \"machine-config-controller-84d6567774-t68ql\" (UID: \"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.872640 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h5xn\" (UniqueName: \"kubernetes.io/projected/963f6709-e807-40bb-aa9e-9bba24e82b71-kube-api-access-7h5xn\") pod \"cluster-image-registry-operator-dc59b4c8b-v9m24\" (UID: \"963f6709-e807-40bb-aa9e-9bba24e82b71\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.897339 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.911556 4907 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.917747 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.929139 4907 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.937416 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.947013 4907 projected.go:288] Couldn't get configMap openshift-route-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: E0313 14:09:02.969283 4907 projected.go:288] Couldn't get configMap openshift-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:02 crc kubenswrapper[4907]: I0313 14:09:02.996097 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.016627 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.039597 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.040127 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rnh5t"] Mar 13 14:09:03 crc kubenswrapper[4907]: W0313 14:09:03.047451 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3baeb4f_988d_4d42_b469_a310ca127745.slice/crio-c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04 WatchSource:0}: Error finding container c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04: Status 404 returned error can't find the container with id c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04 Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.058331 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.075993 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59980734-2cf7-42dc-9f6f-a6de7e1c3665-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076499 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-default-certificate\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-config\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-serving-cert\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076560 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e8bf08-339f-4948-9fe2-ab4092d9a942-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076651 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59980734-2cf7-42dc-9f6f-a6de7e1c3665-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076732 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076796 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.076845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-config\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077119 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-trusted-ca\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077144 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-stats-auth\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-service-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077218 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-serving-cert\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077246 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-metrics-certs\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077365 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcq96\" (UniqueName: \"kubernetes.io/projected/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-kube-api-access-vcq96\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c26ccc9f-e682-4b94-9539-cb2607725f71-service-ca-bundle\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077547 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwfx4\" (UniqueName: \"kubernetes.io/projected/699da377-c61b-48f0-92e0-42bd247ee363-kube-api-access-xwfx4\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077611 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d6cj\" (UniqueName: \"kubernetes.io/projected/59980734-2cf7-42dc-9f6f-a6de7e1c3665-kube-api-access-5d6cj\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077654 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077771 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077822 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6bl2\" (UniqueName: \"kubernetes.io/projected/c26ccc9f-e682-4b94-9539-cb2607725f71-kube-api-access-q6bl2\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077895 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebd59055-333a-46c2-ba30-826296414a4c-metrics-tls\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.077975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-serving-cert\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078023 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-etcd-client\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078058 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ctrn\" (UniqueName: \"kubernetes.io/projected/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-kube-api-access-8ctrn\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078103 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6nzw\" (UniqueName: \"kubernetes.io/projected/f8ea4f24-05b0-4661-a617-a4208f9a9188-kube-api-access-j6nzw\") pod \"downloads-7954f5f757-544df\" (UID: \"f8ea4f24-05b0-4661-a617-a4208f9a9188\") " pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078161 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e8bf08-339f-4948-9fe2-ab4092d9a942-config\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10e8bf08-339f-4948-9fe2-ab4092d9a942-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078212 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0396c426-1d56-44e6-885a-1b97d69ad57a-config\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078239 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078318 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.078559 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0396c426-1d56-44e6-885a-1b97d69ad57a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.078689 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.578669328 +0000 UTC m=+242.478457167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.096740 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.116276 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.135293 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.156463 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.176050 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179283 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.179454 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.679417877 +0000 UTC m=+242.579205586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7e89f101-cb61-4117-841d-58b5cc2bdd80-tmpfs\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179604 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-auth-proxy-config\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c934a2b9-3585-4491-8962-df085c473d90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0396c426-1d56-44e6-885a-1b97d69ad57a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.179820 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfn75\" (UniqueName: \"kubernetes.io/projected/c934a2b9-3585-4491-8962-df085c473d90-kube-api-access-lfn75\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.179872 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.679856608 +0000 UTC m=+242.579644317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180149 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1954f507-7cb0-4942-a4c8-6586d370e717-signing-cabundle\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180177 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180256 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgp25\" (UniqueName: \"kubernetes.io/projected/4682bbe1-4257-4d04-a6d7-60c4655d6873-kube-api-access-bgp25\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180346 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59980734-2cf7-42dc-9f6f-a6de7e1c3665-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-config\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180462 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-default-certificate\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnbr7\" (UniqueName: \"kubernetes.io/projected/7810633c-9d54-4cbe-8861-b523f724ec9e-kube-api-access-hnbr7\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-srv-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180626 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq2kr\" (UniqueName: \"kubernetes.io/projected/1954f507-7cb0-4942-a4c8-6586d370e717-kube-api-access-wq2kr\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180751 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-serving-cert\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.180940 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e8bf08-339f-4948-9fe2-ab4092d9a942-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-socket-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181368 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwlcl\" (UniqueName: \"kubernetes.io/projected/fa265490-c23b-4b05-94bc-78b4a636e065-kube-api-access-dwlcl\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181409 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-webhook-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181432 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/3a661298-271c-4c57-94ea-3d69a27d7ad5-kube-api-access-tc8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.181462 4907 projected.go:194] Error preparing data for projected volume kube-api-access-57msq for pod openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181476 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-csi-data-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181507 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-srv-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181528 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.181538 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq podName:1be912b5-9d1d-48bd-a0ab-7f601f9eb9da nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.681524853 +0000 UTC m=+242.581312562 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-57msq" (UniqueName: "kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq") pod "apiserver-7bbb656c7d-ksgr6" (UID: "1be912b5-9d1d-48bd-a0ab-7f601f9eb9da") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181663 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59980734-2cf7-42dc-9f6f-a6de7e1c3665-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181733 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181767 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-config\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181792 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-trusted-ca\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-stats-auth\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181852 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qllsp\" (UniqueName: \"kubernetes.io/projected/4a8597a4-c72f-4023-b2fb-1b1237411a36-kube-api-access-qllsp\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181889 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-service-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181906 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff0e893f-05cb-4751-9960-281c58000519-config\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-serving-cert\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181953 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx6xs\" (UniqueName: \"kubernetes.io/projected/8013cba2-6b3c-4548-af5c-6dd0057da0c7-kube-api-access-lx6xs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.181970 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-node-bootstrap-token\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.182061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.182232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-metrics-certs\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.182521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhdhq\" (UniqueName: \"kubernetes.io/projected/ff214b86-eb3d-46c6-b45a-18a9f63e742c-kube-api-access-zhdhq\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183462 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcq96\" (UniqueName: \"kubernetes.io/projected/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-kube-api-access-vcq96\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183503 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-plugins-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.183521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.184079 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.184118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.184169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c26ccc9f-e682-4b94-9539-cb2607725f71-service-ca-bundle\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnmnd\" (UniqueName: \"kubernetes.io/projected/814322e3-6251-4ea3-a86b-7b89ca6ea728-kube-api-access-gnmnd\") pod \"migrator-59844c95c7-vfbsx\" (UID: \"814322e3-6251-4ea3-a86b-7b89ca6ea728\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185246 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pzkh\" (UniqueName: \"kubernetes.io/projected/7e89f101-cb61-4117-841d-58b5cc2bdd80-kube-api-access-9pzkh\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185316 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-certs\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185492 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185538 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185588 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185636 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwfx4\" (UniqueName: \"kubernetes.io/projected/699da377-c61b-48f0-92e0-42bd247ee363-kube-api-access-xwfx4\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185692 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d6cj\" (UniqueName: \"kubernetes.io/projected/59980734-2cf7-42dc-9f6f-a6de7e1c3665-kube-api-access-5d6cj\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185742 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185916 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff214b86-eb3d-46c6-b45a-18a9f63e742c-cert\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.185993 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186028 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"auto-csr-approver-29556848-v8kv4\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c934a2b9-3585-4491-8962-df085c473d90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186142 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5x9f\" (UniqueName: \"kubernetes.io/projected/ff0e893f-05cb-4751-9960-281c58000519-kube-api-access-n5x9f\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186200 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6bl2\" (UniqueName: \"kubernetes.io/projected/c26ccc9f-e682-4b94-9539-cb2607725f71-kube-api-access-q6bl2\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186284 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fa265490-c23b-4b05-94bc-78b4a636e065-metrics-tls\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-mountpoint-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186369 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qzhc\" (UniqueName: \"kubernetes.io/projected/74b29e18-7d73-4ef5-9739-04137dd1b191-kube-api-access-9qzhc\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.186368 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187200 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187247 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa265490-c23b-4b05-94bc-78b4a636e065-config-volume\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187276 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-profile-collector-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187329 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1954f507-7cb0-4942-a4c8-6586d370e717-signing-key\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebd59055-333a-46c2-ba30-826296414a4c-metrics-tls\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.187373 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-serving-cert\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.188340 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ctrn\" (UniqueName: \"kubernetes.io/projected/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-kube-api-access-8ctrn\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.188393 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-etcd-client\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-registration-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6nzw\" (UniqueName: \"kubernetes.io/projected/f8ea4f24-05b0-4661-a617-a4208f9a9188-kube-api-access-j6nzw\") pod \"downloads-7954f5f757-544df\" (UID: \"f8ea4f24-05b0-4661-a617-a4208f9a9188\") " pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189493 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff0e893f-05cb-4751-9960-281c58000519-serving-cert\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.189837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e8bf08-339f-4948-9fe2-ab4092d9a942-config\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190086 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10e8bf08-339f-4948-9fe2-ab4092d9a942-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0396c426-1d56-44e6-885a-1b97d69ad57a-config\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190210 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtbj8\" (UniqueName: \"kubernetes.io/projected/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-kube-api-access-dtbj8\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190513 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwz8\" (UniqueName: \"kubernetes.io/projected/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-kube-api-access-jgwz8\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.190708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.192627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.196051 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.198780 4907 projected.go:194] Error preparing data for projected volume kube-api-access-mtqz9 for pod openshift-authentication/oauth-openshift-558db77b4-ng4kz: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.198843 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9 podName:3661f6b2-1d5e-42f0-947f-a32b66258a1f nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.698826374 +0000 UTC m=+242.598614063 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mtqz9" (UniqueName: "kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9") pod "oauth-openshift-558db77b4-ng4kz" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.216228 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.219951 4907 projected.go:194] Error preparing data for projected volume kube-api-access-hlg49 for pod openshift-apiserver/apiserver-76f77b778f-r86r4: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.220072 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49 podName:5a9c889b-5954-4d3a-ae72-12000ee1cc25 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.720043801 +0000 UTC m=+242.619831530 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hlg49" (UniqueName: "kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49") pod "apiserver-76f77b778f-r86r4" (UID: "5a9c889b-5954-4d3a-ae72-12000ee1cc25") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.221822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" event={"ID":"a3baeb4f-988d-4d42-b469-a310ca127745","Type":"ContainerStarted","Data":"c8e2efcf239dc9215a99f09f489a8e23320bb5126f2409f2a08ac75994d91c04"} Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.236205 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.256176 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.263961 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9d44a1e-0eab-4bd6-833f-143436e43758-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-hh9sk\" (UID: \"e9d44a1e-0eab-4bd6-833f-143436e43758\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.278519 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.291906 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.292044 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.792022488 +0000 UTC m=+242.691810187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnbr7\" (UniqueName: \"kubernetes.io/projected/7810633c-9d54-4cbe-8861-b523f724ec9e-kube-api-access-hnbr7\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-srv-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292242 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq2kr\" (UniqueName: \"kubernetes.io/projected/1954f507-7cb0-4942-a4c8-6586d370e717-kube-api-access-wq2kr\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292310 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-socket-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-webhook-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292357 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/3a661298-271c-4c57-94ea-3d69a27d7ad5-kube-api-access-tc8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwlcl\" (UniqueName: \"kubernetes.io/projected/fa265490-c23b-4b05-94bc-78b4a636e065-kube-api-access-dwlcl\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292421 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-srv-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292443 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-csi-data-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292522 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qllsp\" (UniqueName: \"kubernetes.io/projected/4a8597a4-c72f-4023-b2fb-1b1237411a36-kube-api-access-qllsp\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292552 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff0e893f-05cb-4751-9960-281c58000519-config\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx6xs\" (UniqueName: \"kubernetes.io/projected/8013cba2-6b3c-4548-af5c-6dd0057da0c7-kube-api-access-lx6xs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-node-bootstrap-token\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292639 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhdhq\" (UniqueName: \"kubernetes.io/projected/ff214b86-eb3d-46c6-b45a-18a9f63e742c-kube-api-access-zhdhq\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292711 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-plugins-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292735 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292760 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292819 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnmnd\" (UniqueName: \"kubernetes.io/projected/814322e3-6251-4ea3-a86b-7b89ca6ea728-kube-api-access-gnmnd\") pod \"migrator-59844c95c7-vfbsx\" (UID: \"814322e3-6251-4ea3-a86b-7b89ca6ea728\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292851 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pzkh\" (UniqueName: \"kubernetes.io/projected/7e89f101-cb61-4117-841d-58b5cc2bdd80-kube-api-access-9pzkh\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292929 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-certs\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.292982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293014 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293100 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-socket-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293117 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff214b86-eb3d-46c6-b45a-18a9f63e742c-cert\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293142 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c934a2b9-3585-4491-8962-df085c473d90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293177 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"auto-csr-approver-29556848-v8kv4\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293203 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5x9f\" (UniqueName: \"kubernetes.io/projected/ff0e893f-05cb-4751-9960-281c58000519-kube-api-access-n5x9f\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293235 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293258 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fa265490-c23b-4b05-94bc-78b4a636e065-metrics-tls\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293293 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-mountpoint-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qzhc\" (UniqueName: \"kubernetes.io/projected/74b29e18-7d73-4ef5-9739-04137dd1b191-kube-api-access-9qzhc\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293381 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa265490-c23b-4b05-94bc-78b4a636e065-config-volume\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293414 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293456 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-profile-collector-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1954f507-7cb0-4942-a4c8-6586d370e717-signing-key\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293562 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-registration-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293634 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff0e893f-05cb-4751-9960-281c58000519-serving-cert\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtbj8\" (UniqueName: \"kubernetes.io/projected/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-kube-api-access-dtbj8\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293719 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-plugins-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293819 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgwz8\" (UniqueName: \"kubernetes.io/projected/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-kube-api-access-jgwz8\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293860 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7e89f101-cb61-4117-841d-58b5cc2bdd80-tmpfs\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-csi-data-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-auth-proxy-config\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293993 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c934a2b9-3585-4491-8962-df085c473d90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfn75\" (UniqueName: \"kubernetes.io/projected/c934a2b9-3585-4491-8962-df085c473d90-kube-api-access-lfn75\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1954f507-7cb0-4942-a4c8-6586d370e717-signing-cabundle\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294067 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff0e893f-05cb-4751-9960-281c58000519-config\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294093 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgp25\" (UniqueName: \"kubernetes.io/projected/4682bbe1-4257-4d04-a6d7-60c4655d6873-kube-api-access-bgp25\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.294271 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.794259359 +0000 UTC m=+242.694047048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7e89f101-cb61-4117-841d-58b5cc2bdd80-tmpfs\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.294833 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c934a2b9-3585-4491-8962-df085c473d90-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.295320 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-auth-proxy-config\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.293724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.296104 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-node-bootstrap-token\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.296806 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1954f507-7cb0-4942-a4c8-6586d370e717-signing-cabundle\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.296820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-apiservice-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.297906 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.298286 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c934a2b9-3585-4491-8962-df085c473d90-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.298938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-profile-collector-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299392 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299583 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7810633c-9d54-4cbe-8861-b523f724ec9e-srv-cert\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-mountpoint-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.299949 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4a8597a4-c72f-4023-b2fb-1b1237411a36-registration-dir\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.300007 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-profile-collector-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.300068 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa265490-c23b-4b05-94bc-78b4a636e065-config-volume\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.300382 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4682bbe1-4257-4d04-a6d7-60c4655d6873-srv-cert\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301067 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-certs\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301285 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff214b86-eb3d-46c6-b45a-18a9f63e742c-cert\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fa265490-c23b-4b05-94bc-78b4a636e065-metrics-tls\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7e89f101-cb61-4117-841d-58b5cc2bdd80-webhook-cert\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.301994 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ff0e893f-05cb-4751-9960-281c58000519-serving-cert\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.302127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.302316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1954f507-7cb0-4942-a4c8-6586d370e717-signing-key\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.315745 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.321949 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s47gg for pod openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.322043 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg podName:71c234e9-5f3d-4b0c-94d4-80b4e33a6dda nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.822020164 +0000 UTC m=+242.721807943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s47gg" (UniqueName: "kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg") pod "cluster-samples-operator-665b6dd947-wjrms" (UID: "71c234e9-5f3d-4b0c-94d4-80b4e33a6dda") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.335690 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.340107 4907 projected.go:194] Error preparing data for projected volume kube-api-access-9qt98 for pod openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.340180 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98 podName:005dfe54-e402-4aa4-8b67-f7e7b685debd nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.840157947 +0000 UTC m=+242.739945706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9qt98" (UniqueName: "kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98") pod "machine-approver-56656f9798-52vhb" (UID: "005dfe54-e402-4aa4-8b67-f7e7b685debd") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.355642 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.358440 4907 projected.go:194] Error preparing data for projected volume kube-api-access-gqtv7 for pod openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.358566 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7 podName:9bb1952d-0e55-4e45-81c0-69a8cddb827e nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.858537926 +0000 UTC m=+242.758325655 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gqtv7" (UniqueName: "kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7") pod "route-controller-manager-6576b87f9c-4h9q7" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.376606 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.379631 4907 projected.go:194] Error preparing data for projected volume kube-api-access-gn5jt for pod openshift-controller-manager/controller-manager-879f6c89f-kwfhg: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.379741 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt podName:814c6dec-6e96-4068-9aec-9e57fefca5db nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.879710352 +0000 UTC m=+242.779498081 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gn5jt" (UniqueName: "kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt") pod "controller-manager-879f6c89f-kwfhg" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.395539 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.395663 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.895648105 +0000 UTC m=+242.795435784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.396191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.396845 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.896827497 +0000 UTC m=+242.796615226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.416362 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.417422 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.435686 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.436606 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.455743 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.463055 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jwmk\" (UniqueName: \"kubernetes.io/projected/b75dbfe3-8887-4a0b-9541-a4e4000924cb-kube-api-access-7jwmk\") pod \"machine-api-operator-5694c8668f-prlhd\" (UID: \"b75dbfe3-8887-4a0b-9541-a4e4000924cb\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.476456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.486355 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0396c426-1d56-44e6-885a-1b97d69ad57a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.496206 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.498791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.499015 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:03.998990264 +0000 UTC m=+242.898777973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.499834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.500375 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.000359592 +0000 UTC m=+242.900147311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.504441 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59980734-2cf7-42dc-9f6f-a6de7e1c3665-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.516306 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.521420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-config\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.538608 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.546367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-default-certificate\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.556567 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.564239 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-serving-cert\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.578080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.585144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.601513 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.601712 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.101677736 +0000 UTC m=+243.001465425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.602182 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.602525 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.102512319 +0000 UTC m=+243.002300008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.616082 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.625153 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10e8bf08-339f-4948-9fe2-ab4092d9a942-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.655372 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.662558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-service-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.675213 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.685546 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59980734-2cf7-42dc-9f6f-a6de7e1c3665-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.696386 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.702671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.702807 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.202786655 +0000 UTC m=+243.102574344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.702987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.703103 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.703337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/699da377-c61b-48f0-92e0-42bd247ee363-etcd-ca\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.703394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.703655 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.203638508 +0000 UTC m=+243.103426197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.706260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57msq\" (UniqueName: \"kubernetes.io/projected/1be912b5-9d1d-48bd-a0ab-7f601f9eb9da-kube-api-access-57msq\") pod \"apiserver-7bbb656c7d-ksgr6\" (UID: \"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.706696 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"oauth-openshift-558db77b4-ng4kz\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.716214 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.728077 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-serving-cert\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.743557 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.753944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-trusted-ca\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.755496 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.766195 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-stats-auth\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.776154 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.783067 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-config\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.796317 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.804972 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.805124 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.305087447 +0000 UTC m=+243.204875136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.805375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.805683 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.806112 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.306095553 +0000 UTC m=+243.205883242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.809875 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c26ccc9f-e682-4b94-9539-cb2607725f71-metrics-certs\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.813261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlg49\" (UniqueName: \"kubernetes.io/projected/5a9c889b-5954-4d3a-ae72-12000ee1cc25-kube-api-access-hlg49\") pod \"apiserver-76f77b778f-r86r4\" (UID: \"5a9c889b-5954-4d3a-ae72-12000ee1cc25\") " pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.837307 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.838942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24"] Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.839946 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql"] Mar 13 14:09:03 crc kubenswrapper[4907]: W0313 14:09:03.845239 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod963f6709_e807_40bb_aa9e_9bba24e82b71.slice/crio-cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837 WatchSource:0}: Error finding container cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837: Status 404 returned error can't find the container with id cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837 Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.845664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c26ccc9f-e682-4b94-9539-cb2607725f71-service-ca-bundle\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.906366 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.906520 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.406503883 +0000 UTC m=+243.306291572 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.906655 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.907272 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.907591 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.907958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.908113 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:03 crc kubenswrapper[4907]: E0313 14:09:03.908300 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.408282671 +0000 UTC m=+243.308070410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910561 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"route-controller-manager-6576b87f9c-4h9q7\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910620 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"controller-manager-879f6c89f-kwfhg\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910631 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s47gg\" (UniqueName: \"kubernetes.io/projected/71c234e9-5f3d-4b0c-94d4-80b4e33a6dda-kube-api-access-s47gg\") pod \"cluster-samples-operator-665b6dd947-wjrms\" (UID: \"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.910862 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.912803 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qt98\" (UniqueName: \"kubernetes.io/projected/005dfe54-e402-4aa4-8b67-f7e7b685debd-kube-api-access-9qt98\") pod \"machine-approver-56656f9798-52vhb\" (UID: \"005dfe54-e402-4aa4-8b67-f7e7b685debd\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.936594 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.943634 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-serving-cert\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.955362 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.964633 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ebd59055-333a-46c2-ba30-826296414a4c-metrics-tls\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.975047 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 14:09:03 crc kubenswrapper[4907]: I0313 14:09:03.982609 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.009695 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.009834 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.509817452 +0000 UTC m=+243.409605141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.010079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.010413 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.510399358 +0000 UTC m=+243.410187057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.015740 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.031445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/699da377-c61b-48f0-92e0-42bd247ee363-etcd-client\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.070016 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.095741 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.102566 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10e8bf08-339f-4948-9fe2-ab4092d9a942-config\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.112231 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.112430 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.612391051 +0000 UTC m=+243.512178760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.112844 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.113218 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.613201493 +0000 UTC m=+243.512989202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.116528 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.121575 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0396c426-1d56-44e6-885a-1b97d69ad57a-config\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.135937 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.146131 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dslzm\" (UniqueName: \"kubernetes.io/projected/ae320273-06e8-43c5-a64f-acd80ad16d1c-kube-api-access-dslzm\") pod \"ingress-operator-5b745b69d9-5szhc\" (UID: \"ae320273-06e8-43c5-a64f-acd80ad16d1c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.155965 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.167558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnmbh\" (UniqueName: \"kubernetes.io/projected/a7b66665-221d-4334-8263-aeb3b20e1c92-kube-api-access-rnmbh\") pod \"openshift-controller-manager-operator-756b6f6bc6-hfcsw\" (UID: \"a7b66665-221d-4334-8263-aeb3b20e1c92\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.195326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnbr7\" (UniqueName: \"kubernetes.io/projected/7810633c-9d54-4cbe-8861-b523f724ec9e-kube-api-access-hnbr7\") pod \"catalog-operator-68c6474976-fwcwx\" (UID: \"7810633c-9d54-4cbe-8861-b523f724ec9e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.213259 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq2kr\" (UniqueName: \"kubernetes.io/projected/1954f507-7cb0-4942-a4c8-6586d370e717-kube-api-access-wq2kr\") pod \"service-ca-9c57cc56f-zwzc4\" (UID: \"1954f507-7cb0-4942-a4c8-6586d370e717\") " pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.213871 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.214141 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.714111626 +0000 UTC m=+243.613899355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.214648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.215060 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.715042742 +0000 UTC m=+243.614830501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.228484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" event={"ID":"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c","Type":"ContainerStarted","Data":"f7123489a4da754753944b4f46a995a572d6483f103454fc1d99c5f31f23fc56"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.228531 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" event={"ID":"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c","Type":"ContainerStarted","Data":"0c7c592b44c27e210c582853669a9bdfbaaf66d54ff54f806a5a7d997b04aaff"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.228545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" event={"ID":"a8e9ea4a-43ad-4da0-99fa-c01eea712c8c","Type":"ContainerStarted","Data":"7becc3bfa86aaf701a6ac61da68318dbe069d09afac1a898ddc8d52a3fe71572"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.230308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" event={"ID":"a3baeb4f-988d-4d42-b469-a310ca127745","Type":"ContainerStarted","Data":"093846580b2ef04d12bd65ee80e3c459927d2923b16915bfe3e4a7dabe9c4e6a"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.231985 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" event={"ID":"963f6709-e807-40bb-aa9e-9bba24e82b71","Type":"ContainerStarted","Data":"227dc9775ef9c7db63ec63427d557314bc07d97ac6e9b074d22c6fedfee02153"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.232012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" event={"ID":"963f6709-e807-40bb-aa9e-9bba24e82b71","Type":"ContainerStarted","Data":"cfdd9cd399e1567ff60aac5628b25525ccb0e61a814a25a3fc9df094f6cb8837"} Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.238856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc8b2\" (UniqueName: \"kubernetes.io/projected/3a661298-271c-4c57-94ea-3d69a27d7ad5-kube-api-access-tc8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.251268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"auto-csr-approver-29556848-v8kv4\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.274874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pzkh\" (UniqueName: \"kubernetes.io/projected/7e89f101-cb61-4117-841d-58b5cc2bdd80-kube-api-access-9pzkh\") pod \"packageserver-d55dfcdfc-pk65f\" (UID: \"7e89f101-cb61-4117-841d-58b5cc2bdd80\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.294133 4907 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.294245 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls podName:3a661298-271c-4c57-94ea-3d69a27d7ad5 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.794219354 +0000 UTC m=+243.694007113 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-djph7" (UID: "3a661298-271c-4c57-94ea-3d69a27d7ad5") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.294163 4907 request.go:700] Waited for 1.000276454s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/serviceaccounts/marketplace-operator/token Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.296471 4907 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.296529 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images podName:74b29e18-7d73-4ef5-9739-04137dd1b191 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.796517747 +0000 UTC m=+243.696305436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images") pod "machine-config-operator-74547568cd-92gs7" (UID: "74b29e18-7d73-4ef5-9739-04137dd1b191") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300321 4907 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300353 4907 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300399 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls podName:74b29e18-7d73-4ef5-9739-04137dd1b191 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.800383081 +0000 UTC m=+243.700170770 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls") pod "machine-config-operator-74547568cd-92gs7" (UID: "74b29e18-7d73-4ef5-9739-04137dd1b191") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.300419 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs podName:8013cba2-6b3c-4548-af5c-6dd0057da0c7 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.800412532 +0000 UTC m=+243.700200221 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs") pod "multus-admission-controller-857f4d67dd-ksfb8" (UID: "8013cba2-6b3c-4548-af5c-6dd0057da0c7") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.301818 4907 projected.go:288] Couldn't get configMap openshift-console/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.301861 4907 projected.go:194] Error preparing data for projected volume kube-api-access-vxlx4 for pod openshift-console/console-f9d7485db-l9xpb: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.301968 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4 podName:4c8c0266-5452-4abb-92c5-c536df94dd41 nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.801939144 +0000 UTC m=+243.701726903 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vxlx4" (UniqueName: "kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4") pod "console-f9d7485db-l9xpb" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.304066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.314450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"marketplace-operator-79b997595-5stnc\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.315982 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.316295 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.816266614 +0000 UTC m=+243.716054343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.317792 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.318394 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.818380231 +0000 UTC m=+243.718167920 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.331258 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5x9f\" (UniqueName: \"kubernetes.io/projected/ff0e893f-05cb-4751-9960-281c58000519-kube-api-access-n5x9f\") pod \"service-ca-operator-777779d784-75gdt\" (UID: \"ff0e893f-05cb-4751-9960-281c58000519\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.331631 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.353644 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.363664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwlcl\" (UniqueName: \"kubernetes.io/projected/fa265490-c23b-4b05-94bc-78b4a636e065-kube-api-access-dwlcl\") pod \"dns-default-8jqgq\" (UID: \"fa265490-c23b-4b05-94bc-78b4a636e065\") " pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.371196 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qllsp\" (UniqueName: \"kubernetes.io/projected/4a8597a4-c72f-4023-b2fb-1b1237411a36-kube-api-access-qllsp\") pod \"csi-hostpathplugin-sjwvn\" (UID: \"4a8597a4-c72f-4023-b2fb-1b1237411a36\") " pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.376030 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.376087 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.379639 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.389990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.397937 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhdhq\" (UniqueName: \"kubernetes.io/projected/ff214b86-eb3d-46c6-b45a-18a9f63e742c-kube-api-access-zhdhq\") pod \"ingress-canary-252nq\" (UID: \"ff214b86-eb3d-46c6-b45a-18a9f63e742c\") " pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.417502 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.418596 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.418742 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.918720909 +0000 UTC m=+243.818508608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.418968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.419489 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:04.91947299 +0000 UTC m=+243.819260689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.425271 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgp25\" (UniqueName: \"kubernetes.io/projected/4682bbe1-4257-4d04-a6d7-60c4655d6873-kube-api-access-bgp25\") pod \"olm-operator-6b444d44fb-hkbqt\" (UID: \"4682bbe1-4257-4d04-a6d7-60c4655d6873\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.430985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-252nq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.446775 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.461551 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgwz8\" (UniqueName: \"kubernetes.io/projected/c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d-kube-api-access-jgwz8\") pod \"machine-config-server-g6q2r\" (UID: \"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d\") " pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.479710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.481922 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.482152 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.482181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.484413 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx6xs\" (UniqueName: \"kubernetes.io/projected/8013cba2-6b3c-4548-af5c-6dd0057da0c7-kube-api-access-lx6xs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.513765 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfn75\" (UniqueName: \"kubernetes.io/projected/c934a2b9-3585-4491-8962-df085c473d90-kube-api-access-lfn75\") pod \"openshift-apiserver-operator-796bbdcf4f-xnr2b\" (UID: \"c934a2b9-3585-4491-8962-df085c473d90\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.518591 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.521393 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.521550 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.021528063 +0000 UTC m=+243.921315752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.522324 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.522560 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.022551962 +0000 UTC m=+243.922339651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.535150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qzhc\" (UniqueName: \"kubernetes.io/projected/74b29e18-7d73-4ef5-9739-04137dd1b191-kube-api-access-9qzhc\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: W0313 14:09:04.535972 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e89f101_cb61_4117_841d_58b5cc2bdd80.slice/crio-5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a WatchSource:0}: Error finding container 5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a: Status 404 returned error can't find the container with id 5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.559491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"collect-profiles-29556840-pgmwj\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.575645 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.580028 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtbj8\" (UniqueName: \"kubernetes.io/projected/b651a92c-e9f6-4a1a-8c49-b6500bb7c61c-kube-api-access-dtbj8\") pod \"package-server-manager-789f6589d5-ccjsv\" (UID: \"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.588820 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.598349 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.601490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.607301 4907 projected.go:288] Couldn't get configMap openshift-kube-apiserver-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.607331 4907 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.607384 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access podName:0396c426-1d56-44e6-885a-1b97d69ad57a nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.107366318 +0000 UTC m=+244.007154007 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access") pod "kube-apiserver-operator-766d6c64bb-rmkfg" (UID: "0396c426-1d56-44e6-885a-1b97d69ad57a") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.615655 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.616471 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.621983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.623494 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.623931 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.123912758 +0000 UTC m=+244.023700447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.626730 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6nzw\" (UniqueName: \"kubernetes.io/projected/f8ea4f24-05b0-4661-a617-a4208f9a9188-kube-api-access-j6nzw\") pod \"downloads-7954f5f757-544df\" (UID: \"f8ea4f24-05b0-4661-a617-a4208f9a9188\") " pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.640281 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.642191 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.647933 4907 projected.go:288] Couldn't get configMap openshift-dns-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.677177 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.697701 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.699559 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.715158 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g6q2r" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.717286 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.735686 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.736291 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.236272822 +0000 UTC m=+244.136060511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.738420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.744574 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.796554 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.799554 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.799581 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.813114 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.835310 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836332 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836516 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836681 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836717 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.836799 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.336760274 +0000 UTC m=+244.236547963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836931 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.836977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.839306 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/74b29e18-7d73-4ef5-9739-04137dd1b191-images\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.848265 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"console-f9d7485db-l9xpb\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.849072 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/74b29e18-7d73-4ef5-9739-04137dd1b191-proxy-tls\") pod \"machine-config-operator-74547568cd-92gs7\" (UID: \"74b29e18-7d73-4ef5-9739-04137dd1b191\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.855147 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8013cba2-6b3c-4548-af5c-6dd0057da0c7-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ksfb8\" (UID: \"8013cba2-6b3c-4548-af5c-6dd0057da0c7\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.855303 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/3a661298-271c-4c57-94ea-3d69a27d7ad5-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-djph7\" (UID: \"3a661298-271c-4c57-94ea-3d69a27d7ad5\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.856506 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.865845 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.880692 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.888210 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.888553 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.900110 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.915924 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.917637 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.938039 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.938867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:04 crc kubenswrapper[4907]: E0313 14:09:04.941562 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.441547652 +0000 UTC m=+244.341335341 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.952389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/10e8bf08-339f-4948-9fe2-ab4092d9a942-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-zz77x\" (UID: \"10e8bf08-339f-4948-9fe2-ab4092d9a942\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.956348 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.965370 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.976280 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.978362 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.980698 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-75gdt"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.983974 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.986320 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:09:04 crc kubenswrapper[4907]: I0313 14:09:04.995958 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.004183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.016311 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.035954 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.043015 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.043431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.043898 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.543858254 +0000 UTC m=+244.443645933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.046653 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.065154 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.068621 4907 projected.go:194] Error preparing data for projected volume kube-api-access-wk2bb for pod openshift-dns-operator/dns-operator-744455d44c-xkdjh: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.068745 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb podName:ebd59055-333a-46c2-ba30-826296414a4c nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.568718 +0000 UTC m=+244.468505689 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-wk2bb" (UniqueName: "kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb") pod "dns-operator-744455d44c-xkdjh" (UID: "ebd59055-333a-46c2-ba30-826296414a4c") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.071325 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-zwzc4"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.085652 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-prlhd"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.090372 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-252nq"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.096490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.102191 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-sjwvn"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.104301 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.116136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.121008 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcq96\" (UniqueName: \"kubernetes.io/projected/9b75954c-b75d-4f17-82dc-4c0358fd8d0f-kube-api-access-vcq96\") pod \"console-operator-58897d9998-w76kw\" (UID: \"9b75954c-b75d-4f17-82dc-4c0358fd8d0f\") " pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.140939 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.141240 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9d44a1e_0eab_4bd6_833f_143436e43758.slice/crio-23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93 WatchSource:0}: Error finding container 23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93: Status 404 returned error can't find the container with id 23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.147004 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6bl2\" (UniqueName: \"kubernetes.io/projected/c26ccc9f-e682-4b94-9539-cb2607725f71-kube-api-access-q6bl2\") pod \"router-default-5444994796-f8mmz\" (UID: \"c26ccc9f-e682-4b94-9539-cb2607725f71\") " pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.147005 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.147169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.147277 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.647264335 +0000 UTC m=+244.547052024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.156521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0396c426-1d56-44e6-885a-1b97d69ad57a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-rmkfg\" (UID: \"0396c426-1d56-44e6-885a-1b97d69ad57a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.158266 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.173749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwfx4\" (UniqueName: \"kubernetes.io/projected/699da377-c61b-48f0-92e0-42bd247ee363-kube-api-access-xwfx4\") pod \"etcd-operator-b45778765-hwf6r\" (UID: \"699da377-c61b-48f0-92e0-42bd247ee363\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.176375 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.195154 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d6cj\" (UniqueName: \"kubernetes.io/projected/59980734-2cf7-42dc-9f6f-a6de7e1c3665-kube-api-access-5d6cj\") pod \"kube-storage-version-migrator-operator-b67b599dd-56rqr\" (UID: \"59980734-2cf7-42dc-9f6f-a6de7e1c3665\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.201573 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.201760 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.216704 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.228759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.229584 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.231799 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ctrn\" (UniqueName: \"kubernetes.io/projected/0149e083-ed5b-4c7f-94e7-4f3f2ff8a557-kube-api-access-8ctrn\") pod \"openshift-config-operator-7777fb866f-wgpdd\" (UID: \"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.245076 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.253275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.253647 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.753632197 +0000 UTC m=+244.653419886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.260495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnmnd\" (UniqueName: \"kubernetes.io/projected/814322e3-6251-4ea3-a86b-7b89ca6ea728-kube-api-access-gnmnd\") pod \"migrator-59844c95c7-vfbsx\" (UID: \"814322e3-6251-4ea3-a86b-7b89ca6ea728\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.266009 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.269529 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.275559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.288720 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8jqgq"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.288776 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.298346 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.301453 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.305760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" event={"ID":"005dfe54-e402-4aa4-8b67-f7e7b685debd","Type":"ContainerStarted","Data":"2640caee6d284f0bc297ee7eb18a0e186b53f36e618ca72302ac3834c8f7aeaa"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.307004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" event={"ID":"a8867781-179d-47e4-9a55-330deb10d273","Type":"ContainerStarted","Data":"38620171197d496fb764e2284a8169d5eff52a1213907791553d488eaad39d9d"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.309290 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.316198 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.316508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.331284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" event={"ID":"7e89f101-cb61-4117-841d-58b5cc2bdd80","Type":"ContainerStarted","Data":"ed7d771c7834263888b56c54ee8c8ea22b90724c2d98378a98feddf4aea2b65b"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.331329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" event={"ID":"7e89f101-cb61-4117-841d-58b5cc2bdd80","Type":"ContainerStarted","Data":"5b8f9375c0aaaf442ecdf173ca5f3a8f53bf0c84929853df0d2803ae5825f27a"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.332434 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.333508 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19edf37b_a1d3_4c7c_945e_5d0b8bb601bf.slice/crio-e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36 WatchSource:0}: Error finding container e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36: Status 404 returned error can't find the container with id e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.337493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" event={"ID":"ff0e893f-05cb-4751-9960-281c58000519","Type":"ContainerStarted","Data":"4508a6e2419581abd5918c50470d548f610ef0c89853506353a8483b8ea3015d"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.340168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.341417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.343182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" event={"ID":"e9d44a1e-0eab-4bd6-833f-143436e43758","Type":"ContainerStarted","Data":"23e4a4539e1b648af539df572b036a3e5ceebfe6280418e6d7f8719a1fd54f93"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.343898 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"a38dcdd81d9bc13e5a8dae158690d917c7091d0f83e3c30429ea70f671e8035f"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.352550 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" event={"ID":"7810633c-9d54-4cbe-8861-b523f724ec9e","Type":"ContainerStarted","Data":"afd7d366e1cbef135eee0552758925fb36fb60a4c43a2c12c80b73605f7cb0fa"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.352701 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" event={"ID":"7810633c-9d54-4cbe-8861-b523f724ec9e","Type":"ContainerStarted","Data":"6ef82b59f33daccea1490dc3bef81544cb46172092caf29fdc1c89470a456854"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.353123 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.354648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.355560 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.855546818 +0000 UTC m=+244.755334507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.364021 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.367604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.368789 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g6q2r" event={"ID":"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d","Type":"ContainerStarted","Data":"f18d9e133ffdf3c70e7a18735c7450c5821e6f17755cb96b4dc3e0af638b395c"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.368820 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g6q2r" event={"ID":"c77fd554-e957-4d49-b8cd-b4f7dcd4ba2d","Type":"ContainerStarted","Data":"9d2d79f538bec30c7e5525cc958a7eb716f40a8b39789138099658c3006687d4"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.376064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-252nq" event={"ID":"ff214b86-eb3d-46c6-b45a-18a9f63e742c","Type":"ContainerStarted","Data":"eeeb42502a81f1b22e077bafdd88ba9c9f5f66a3268d2f62ba58eea461cb1789"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.385011 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" event={"ID":"b75dbfe3-8887-4a0b-9541-a4e4000924cb","Type":"ContainerStarted","Data":"e7d019e94109b87308c200c5fd282b0abdb2116ded6cffd570ca3ed83675f3ad"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.393774 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" event={"ID":"1954f507-7cb0-4942-a4c8-6586d370e717","Type":"ContainerStarted","Data":"b2c639cb1f8fc378273d3d9722547069101a439330919f458fbd4e276e30511f"} Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.398701 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.410019 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.438572 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.446823 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.450659 4907 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pk65f container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.450746 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" podUID="7e89f101-cb61-4117-841d-58b5cc2bdd80" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.455987 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.456431 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:05.95640496 +0000 UTC m=+244.856192649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.459308 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.459408 4907 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fwcwx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.459471 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" podUID="7810633c-9d54-4cbe-8861-b523f724ec9e" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.460532 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.474585 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3661f6b2_1d5e_42f0_947f_a32b66258a1f.slice/crio-9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506 WatchSource:0}: Error finding container 9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506: Status 404 returned error can't find the container with id 9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.476943 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.484325 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.517907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.522982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.536104 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.545724 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.558267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.558581 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.058570078 +0000 UTC m=+244.958357767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.585295 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.597077 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-r86r4"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.597342 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.663170 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.663326 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.664902 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.164857447 +0000 UTC m=+245.064645146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.682802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk2bb\" (UniqueName: \"kubernetes.io/projected/ebd59055-333a-46c2-ba30-826296414a4c-kube-api-access-wk2bb\") pod \"dns-operator-744455d44c-xkdjh\" (UID: \"ebd59055-333a-46c2-ba30-826296414a4c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.735359 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.766600 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.766923 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.266911152 +0000 UTC m=+245.166698841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.771567 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod814c6dec_6e96_4068_9aec_9e57fefca5db.slice/crio-0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5 WatchSource:0}: Error finding container 0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5: Status 404 returned error can't find the container with id 0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.805839 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.810682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ksfb8"] Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.835544 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7b66665_221d_4334_8263_aeb3b20e1c92.slice/crio-bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c WatchSource:0}: Error finding container bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c: Status 404 returned error can't find the container with id bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.862549 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.871444 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.875491 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms"] Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.875751 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.876103 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.3760883 +0000 UTC m=+245.275875989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.908616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-544df"] Mar 13 14:09:05 crc kubenswrapper[4907]: W0313 14:09:05.917029 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a9c889b_5954_4d3a_ae72_12000ee1cc25.slice/crio-23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591 WatchSource:0}: Error finding container 23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591: Status 404 returned error can't find the container with id 23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591 Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.977145 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:05 crc kubenswrapper[4907]: E0313 14:09:05.977457 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.477442675 +0000 UTC m=+245.377230364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:05 crc kubenswrapper[4907]: I0313 14:09:05.998420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.000781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.001941 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x"] Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.002925 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74b29e18_7d73_4ef5_9739_04137dd1b191.slice/crio-28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060 WatchSource:0}: Error finding container 28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060: Status 404 returned error can't find the container with id 28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060 Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.003213 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8013cba2_6b3c_4548_af5c_6dd0057da0c7.slice/crio-bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15 WatchSource:0}: Error finding container bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15: Status 404 returned error can't find the container with id bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15 Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.003730 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bb1952d_0e55_4e45_81c0_69a8cddb827e.slice/crio-9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a WatchSource:0}: Error finding container 9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a: Status 404 returned error can't find the container with id 9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.039361 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.078444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.079118 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.579074099 +0000 UTC m=+245.478861788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.079336 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.079837 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.579826119 +0000 UTC m=+245.479613808 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.120354 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e8bf08_339f_4948_9fe2_ab4092d9a942.slice/crio-20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898 WatchSource:0}: Error finding container 20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898: Status 404 returned error can't find the container with id 20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898 Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.158632 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hwf6r"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.180229 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.180762 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.680748072 +0000 UTC m=+245.580535761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.263508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.281987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.282304 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.782291213 +0000 UTC m=+245.682078902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.327280 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" podStartSLOduration=185.327261406 podStartE2EDuration="3m5.327261406s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.289031847 +0000 UTC m=+245.188819536" watchObservedRunningTime="2026-03-13 14:09:06.327261406 +0000 UTC m=+245.227049095" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.382897 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.383035 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.883018222 +0000 UTC m=+245.782805911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.383080 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.383411 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.883402682 +0000 UTC m=+245.783190361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.410184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" event={"ID":"e9d44a1e-0eab-4bd6-833f-143436e43758","Type":"ContainerStarted","Data":"f2b59b770051ee0c497bccf354ff8d97a4c4e78381941599f2b81bcdb45aea51"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.413796 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerStarted","Data":"9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.415228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" event={"ID":"699da377-c61b-48f0-92e0-42bd247ee363","Type":"ContainerStarted","Data":"98417ee0f0233faf0047977c4ec6593e0198d88401a88c6672ff9dc399e06656"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.417441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"23c50ef680bc0a0105f881d9b6d419e244388a4b1be6b0b744537678408be591"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.418433 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" event={"ID":"8013cba2-6b3c-4548-af5c-6dd0057da0c7","Type":"ContainerStarted","Data":"bed4842855bdf9609aac5b814fc47449456580fd4be12ff6f7dd5cb79033bd15"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.419957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" event={"ID":"a7b66665-221d-4334-8263-aeb3b20e1c92","Type":"ContainerStarted","Data":"bea2a55198cc235ae007e26c1440216d35071366920193590b872d42da99d43c"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.421139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-544df" event={"ID":"f8ea4f24-05b0-4661-a617-a4208f9a9188","Type":"ContainerStarted","Data":"76ec31cdef4dc0d683fc04866509997aaaa312dd52bb20f29c41405f80ba4c5c"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.422272 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerStarted","Data":"97e42fd7291025df6972a91466dcaf06ac63f386d1c5751baba2fa6aca2cf00a"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.426111 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" event={"ID":"1954f507-7cb0-4942-a4c8-6586d370e717","Type":"ContainerStarted","Data":"292d35dc7c24966269168817ed4b7571bbf78be96c1a62a6688c9ddaa28634af"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.433245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" event={"ID":"74b29e18-7d73-4ef5-9739-04137dd1b191","Type":"ContainerStarted","Data":"28e1b4d5a08c50b259a144b90183e591003ab7707068b0fd9883bf559d1a9060"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.441593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" event={"ID":"10e8bf08-339f-4948-9fe2-ab4092d9a942","Type":"ContainerStarted","Data":"20ab56dcb33fadb51a23e57ce832fe3b3f6c0a4e913b6278419b0694bf504898"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.447190 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" event={"ID":"b75dbfe3-8887-4a0b-9541-a4e4000924cb","Type":"ContainerStarted","Data":"bfb81e78d3c0a64712627633cf3de282b68658dfd36fc8d0ea22fe07e87a7ca6"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.448795 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" event={"ID":"c934a2b9-3585-4491-8962-df085c473d90","Type":"ContainerStarted","Data":"e678dddce63a8e965db113d1ad5181b336626e09d4d5b053d64455d2374aa15f"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.449548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerStarted","Data":"2ca75bf0ab6ab7b6d03a36b2b78748709fe1e70503c3e45b40703b8a20dadbca"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.452329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" event={"ID":"4682bbe1-4257-4d04-a6d7-60c4655d6873","Type":"ContainerStarted","Data":"fe97ca6fd77d983d0419febba7976c1afd14125b4f9439aedf1df1df511bf7e0"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.452360 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" event={"ID":"4682bbe1-4257-4d04-a6d7-60c4655d6873","Type":"ContainerStarted","Data":"e3054c999bd2aae22709666dab86d8d7deb5c922a69e6d8a8238d836c3bea89e"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.452907 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.460055 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerStarted","Data":"9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.467034 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" event={"ID":"3a661298-271c-4c57-94ea-3d69a27d7ad5","Type":"ContainerStarted","Data":"d38738b3846a8bae401e7e85068b5989891d2051495c94c078dd0dd338bbc4dd"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.467192 4907 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hkbqt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.467257 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" podUID="4682bbe1-4257-4d04-a6d7-60c4655d6873" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.474478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8jqgq" event={"ID":"fa265490-c23b-4b05-94bc-78b4a636e065","Type":"ContainerStarted","Data":"7915bd087312e5a5b3f4cc701a590ae465651cc0364d1c824e4224d5547bc50b"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.475482 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" event={"ID":"ae320273-06e8-43c5-a64f-acd80ad16d1c","Type":"ContainerStarted","Data":"cb025c3ea8ee29cc007c3e83fffde7d3aa1a71ee99015e66540499acd12ff294"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.479294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" event={"ID":"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da","Type":"ContainerStarted","Data":"917984672ee70e65cbe94b68435b329ada21f7abe45e3abae9a73f7190269e84"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.481395 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerStarted","Data":"0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.486041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.487204 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr"] Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.487523 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:06.987302776 +0000 UTC m=+245.887090465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.509284 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0149e083_ed5b_4c7f_94e7_4f3f2ff8a557.slice/crio-b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262 WatchSource:0}: Error finding container b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262: Status 404 returned error can't find the container with id b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262 Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.517582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerStarted","Data":"d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.517613 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerStarted","Data":"e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.529176 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" event={"ID":"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda","Type":"ContainerStarted","Data":"c587f565b1dce424d527ed3ea9bb069af779f1b90a7481e3460deb763887b3e9"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.534809 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" event={"ID":"005dfe54-e402-4aa4-8b67-f7e7b685debd","Type":"ContainerStarted","Data":"5995450a4779edf85718472df49cceacc642f80ac60d07e668e19501ff70c8b5"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.540852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-252nq" event={"ID":"ff214b86-eb3d-46c6-b45a-18a9f63e742c","Type":"ContainerStarted","Data":"51db44672a2c6c61b21d0063b559e91a222184620152fc66fe9d7966045175a8"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.544490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" event={"ID":"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c","Type":"ContainerStarted","Data":"7ef73955c4af33184c9094d1058ec7a76a57ccb504926b53434a49c80ea3b182"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.556151 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" event={"ID":"ff0e893f-05cb-4751-9960-281c58000519","Type":"ContainerStarted","Data":"8022ae37cda8bf634c08294061821922a69872cde924e7bb3e44901196573e79"} Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.570620 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" podStartSLOduration=185.570600362 podStartE2EDuration="3m5.570600362s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.569252835 +0000 UTC m=+245.469040534" watchObservedRunningTime="2026-03-13 14:09:06.570600362 +0000 UTC m=+245.470388051" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.586917 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fwcwx" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.588685 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.591034 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.091021406 +0000 UTC m=+245.990809085 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.596330 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59980734_2cf7_42dc_9f6f_a6de7e1c3665.slice/crio-1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe WatchSource:0}: Error finding container 1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe: Status 404 returned error can't find the container with id 1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.655402 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-g6q2r" podStartSLOduration=6.655379216 podStartE2EDuration="6.655379216s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.607811643 +0000 UTC m=+245.507599332" watchObservedRunningTime="2026-03-13 14:09:06.655379216 +0000 UTC m=+245.555166905" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.686754 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-t68ql" podStartSLOduration=185.686734939 podStartE2EDuration="3m5.686734939s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.686192734 +0000 UTC m=+245.585980423" watchObservedRunningTime="2026-03-13 14:09:06.686734939 +0000 UTC m=+245.586522628" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.689792 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.690293 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.190272405 +0000 UTC m=+246.090060094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.717402 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.794728 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.795518 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.295505585 +0000 UTC m=+246.195293274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.799044 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xkdjh"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.809799 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w76kw"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.860229 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx"] Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.896175 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.897069 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.397047896 +0000 UTC m=+246.296835585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.961792 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b75954c_b75d_4f17_82dc_4c0358fd8d0f.slice/crio-47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e WatchSource:0}: Error finding container 47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e: Status 404 returned error can't find the container with id 47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e Mar 13 14:09:06 crc kubenswrapper[4907]: W0313 14:09:06.963270 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebd59055_333a_46c2_ba30_826296414a4c.slice/crio-be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c WatchSource:0}: Error finding container be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c: Status 404 returned error can't find the container with id be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.966418 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-v9m24" podStartSLOduration=185.966402382 podStartE2EDuration="3m5.966402382s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:06.927147755 +0000 UTC m=+245.826935434" watchObservedRunningTime="2026-03-13 14:09:06.966402382 +0000 UTC m=+245.866190071" Mar 13 14:09:06 crc kubenswrapper[4907]: I0313 14:09:06.998352 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:06 crc kubenswrapper[4907]: E0313 14:09:06.999270 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.499254075 +0000 UTC m=+246.399041764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.099833 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.100022 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.599995964 +0000 UTC m=+246.499783673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.100368 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.101534 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.601510315 +0000 UTC m=+246.501298004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.137742 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-rnh5t" podStartSLOduration=187.13772344 podStartE2EDuration="3m7.13772344s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.130808112 +0000 UTC m=+246.030595801" watchObservedRunningTime="2026-03-13 14:09:07.13772344 +0000 UTC m=+246.037511119" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.161393 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pk65f" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.202018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.202477 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.702456909 +0000 UTC m=+246.602244598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.309510 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.310447 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.810431905 +0000 UTC m=+246.710219594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.410742 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.411133 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:07.911117542 +0000 UTC m=+246.810905221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.421442 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-252nq" podStartSLOduration=7.421422293 podStartE2EDuration="7.421422293s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.421013161 +0000 UTC m=+246.320800860" watchObservedRunningTime="2026-03-13 14:09:07.421422293 +0000 UTC m=+246.321209982" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.511841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.512519 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.012506949 +0000 UTC m=+246.912294638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.554448 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" podStartSLOduration=186.554423459 podStartE2EDuration="3m6.554423459s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.553925655 +0000 UTC m=+246.453713344" watchObservedRunningTime="2026-03-13 14:09:07.554423459 +0000 UTC m=+246.454211148" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.593823 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerStarted","Data":"40f0b1a25cd0d0a5030ec1dcae203661c52d45d8711eefed49ad300a5495f0ac"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.593867 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerStarted","Data":"b2a81aa0d935e7dfb8b83728724e4149f1365539869eda9b25d8e6618548a262"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.615024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.616155 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.116125976 +0000 UTC m=+247.015913665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.629190 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" event={"ID":"8013cba2-6b3c-4548-af5c-6dd0057da0c7","Type":"ContainerStarted","Data":"fe367c6c7181382e906590402f45585b664a193ada8ba64c3ce22a4ffceb143a"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.636371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" event={"ID":"a7b66665-221d-4334-8263-aeb3b20e1c92","Type":"ContainerStarted","Data":"e725ce3a4600879c951eaa2f5632cd7105e7bd88d5d27083b1e0b8ea856d9f8d"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.664295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" event={"ID":"ebd59055-333a-46c2-ba30-826296414a4c","Type":"ContainerStarted","Data":"be63aa8b685fc3b6de50ab1c5058dba0dd2fe771225810b4038045dc7a41c54c"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.689150 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" podStartSLOduration=186.689118 podStartE2EDuration="3m6.689118s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.686076588 +0000 UTC m=+246.585864277" watchObservedRunningTime="2026-03-13 14:09:07.689118 +0000 UTC m=+246.588905689" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.715031 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-zwzc4" podStartSLOduration=186.715014114 podStartE2EDuration="3m6.715014114s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.7133733 +0000 UTC m=+246.613160989" watchObservedRunningTime="2026-03-13 14:09:07.715014114 +0000 UTC m=+246.614801803" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.717602 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.718685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" event={"ID":"10e8bf08-339f-4948-9fe2-ab4092d9a942","Type":"ContainerStarted","Data":"cb1dbd959b8e29c65b2fa393f724e71e4139cca7b1c49f5b49d06e9a4ef4a6ee"} Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.718747 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.218730435 +0000 UTC m=+247.118518124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.736349 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" event={"ID":"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c","Type":"ContainerStarted","Data":"1060d8d803a2d78da1f8488a9dfc05a98c6bd3ae8f3e202cc535d226fa32fdcb"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.736401 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" event={"ID":"b651a92c-e9f6-4a1a-8c49-b6500bb7c61c","Type":"ContainerStarted","Data":"eee14898537e10df3b0f40a25557131ca5dc13864a11496c9fbd5305b6c03842"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.737088 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.748973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" event={"ID":"0396c426-1d56-44e6-885a-1b97d69ad57a","Type":"ContainerStarted","Data":"a594462937e2e5a805dd6fe9e9b6c749f8349ed77ce762003ee5b78ec459bae8"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.774240 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8jqgq" event={"ID":"fa265490-c23b-4b05-94bc-78b4a636e065","Type":"ContainerStarted","Data":"59bec09f1e914543d24ba435710c730a198b7a8490d1fa1f391f63d54ca5e15b"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.781843 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerStarted","Data":"11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.782512 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.785628 4907 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kwfhg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.785789 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.798517 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" event={"ID":"814322e3-6251-4ea3-a86b-7b89ca6ea728","Type":"ContainerStarted","Data":"c0bccb4ffe867441f6cbe79f037b0841fedef0026ddfa347dd39ec3bd29ad30d"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.800187 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-hh9sk" podStartSLOduration=186.800160369 podStartE2EDuration="3m6.800160369s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.792053198 +0000 UTC m=+246.691840877" watchObservedRunningTime="2026-03-13 14:09:07.800160369 +0000 UTC m=+246.699948058" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.803004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" event={"ID":"b75dbfe3-8887-4a0b-9541-a4e4000924cb","Type":"ContainerStarted","Data":"be13451dfb6a4d9101c5c56519e9742d9a74fed841a8fa3e62ab1471ed23edca"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.805399 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" event={"ID":"c934a2b9-3585-4491-8962-df085c473d90","Type":"ContainerStarted","Data":"78b55d953a014b3c151178d298a92988bb6e4a6dda66d0fb41b8a35f9cd910c8"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.822910 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.824217 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.324191692 +0000 UTC m=+247.223979391 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.840000 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerStarted","Data":"3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.840820 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.842373 4907 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ng4kz container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.842429 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.848567 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"4651f82e390b3a5ac2c55407710eda6955584915c8001c9a1dc4182061851ee7"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.860797 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-75gdt" podStartSLOduration=186.860774807 podStartE2EDuration="3m6.860774807s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.841145003 +0000 UTC m=+246.740932692" watchObservedRunningTime="2026-03-13 14:09:07.860774807 +0000 UTC m=+246.760562496" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.870711 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podStartSLOduration=186.870686496 podStartE2EDuration="3m6.870686496s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.86971706 +0000 UTC m=+246.769504769" watchObservedRunningTime="2026-03-13 14:09:07.870686496 +0000 UTC m=+246.770474195" Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.892911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" event={"ID":"59980734-2cf7-42dc-9f6f-a6de7e1c3665","Type":"ContainerStarted","Data":"1f36e2ae021694a6f49d8cea38eb878db4943e0747adab7e3fe2b00be56274fe"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.898286 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" event={"ID":"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda","Type":"ContainerStarted","Data":"eb3bbb7e5376638446dc533d2a33fce47abb9f3e1fafc7162f893e7aa4800daf"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.925763 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:07 crc kubenswrapper[4907]: E0313 14:09:07.926114 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.426100943 +0000 UTC m=+247.325888632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.964284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" event={"ID":"005dfe54-e402-4aa4-8b67-f7e7b685debd","Type":"ContainerStarted","Data":"9537155f0efd7cadee4facaf87934d2aae807cb49c4f95882646566ae9e4d8a2"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.978035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w76kw" event={"ID":"9b75954c-b75d-4f17-82dc-4c0358fd8d0f","Type":"ContainerStarted","Data":"47724a9f48548516662c355e2e67e0051f4b8870e897b87bf044f8835576d09e"} Mar 13 14:09:07 crc kubenswrapper[4907]: I0313 14:09:07.991474 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xnr2b" podStartSLOduration=187.991457 podStartE2EDuration="3m7.991457s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.961761672 +0000 UTC m=+246.861549361" watchObservedRunningTime="2026-03-13 14:09:07.991457 +0000 UTC m=+246.891244689" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.013672 4907 generic.go:334] "Generic (PLEG): container finished" podID="1be912b5-9d1d-48bd-a0ab-7f601f9eb9da" containerID="9afd4edd5700b10026def68352336f5dc9ba3dd07805b6aa3edb298f50f16142" exitCode=0 Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.013750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" event={"ID":"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da","Type":"ContainerDied","Data":"9afd4edd5700b10026def68352336f5dc9ba3dd07805b6aa3edb298f50f16142"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.027249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.027370 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.527345605 +0000 UTC m=+247.427133294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.027894 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.036199 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-zz77x" podStartSLOduration=187.036180516 podStartE2EDuration="3m7.036180516s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:07.995184391 +0000 UTC m=+246.894972100" watchObservedRunningTime="2026-03-13 14:09:08.036180516 +0000 UTC m=+246.935968205" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.036708 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" podStartSLOduration=187.03670085 podStartE2EDuration="3m7.03670085s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.032414223 +0000 UTC m=+246.932201912" watchObservedRunningTime="2026-03-13 14:09:08.03670085 +0000 UTC m=+246.936488529" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.047143 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.547124553 +0000 UTC m=+247.446912242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.069778 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hfcsw" podStartSLOduration=187.069757539 podStartE2EDuration="3m7.069757539s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.068632808 +0000 UTC m=+246.968420497" watchObservedRunningTime="2026-03-13 14:09:08.069757539 +0000 UTC m=+246.969545228" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.093454 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" event={"ID":"74b29e18-7d73-4ef5-9739-04137dd1b191","Type":"ContainerStarted","Data":"9a979dcc7ab19700343b563444457c87eb78ce1bc410e31bf716d61b5b4c37b4"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.106564 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" event={"ID":"ae320273-06e8-43c5-a64f-acd80ad16d1c","Type":"ContainerStarted","Data":"2e80e8494d9dc6d61c40f9ceab4bc533f542e55642a745e2c701bf307a2db887"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.129867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.135179 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.635147096 +0000 UTC m=+247.534934785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.135663 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerStarted","Data":"48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.138206 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.147117 4907 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5stnc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.147184 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.152326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-544df" event={"ID":"f8ea4f24-05b0-4661-a617-a4208f9a9188","Type":"ContainerStarted","Data":"f48f2ca4ee981d838c5862139a85b9d2f101144d532b7ded20d21d6293a389b6"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.153350 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.162558 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.162636 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.167480 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-prlhd" podStartSLOduration=187.167457224 podStartE2EDuration="3m7.167457224s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.159376245 +0000 UTC m=+247.059163924" watchObservedRunningTime="2026-03-13 14:09:08.167457224 +0000 UTC m=+247.067244913" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.168535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-f8mmz" event={"ID":"c26ccc9f-e682-4b94-9539-cb2607725f71","Type":"ContainerStarted","Data":"4c9b269b13491d5272c98a727d99c52f5328c3fde0c6fcad1345bb2263323a7f"} Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.176314 4907 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-hkbqt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.176375 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" podUID="4682bbe1-4257-4d04-a6d7-60c4655d6873" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.201415 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podStartSLOduration=188.201389088 podStartE2EDuration="3m8.201389088s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.199108805 +0000 UTC m=+247.098896494" watchObservedRunningTime="2026-03-13 14:09:08.201389088 +0000 UTC m=+247.101176777" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.242034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.242394 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" podStartSLOduration=187.242372401 podStartE2EDuration="3m7.242372401s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.241466286 +0000 UTC m=+247.141253975" watchObservedRunningTime="2026-03-13 14:09:08.242372401 +0000 UTC m=+247.142160090" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.244031 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.744013376 +0000 UTC m=+247.643801065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.257956 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.259697 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.262976 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.279727 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.338269 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podStartSLOduration=187.338232097 podStartE2EDuration="3m7.338232097s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.27249705 +0000 UTC m=+247.172284739" watchObservedRunningTime="2026-03-13 14:09:08.338232097 +0000 UTC m=+247.238019796" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343102 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343351 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.343386 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.843365447 +0000 UTC m=+247.743153146 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.343434 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.383611 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-52vhb" podStartSLOduration=188.383587161 podStartE2EDuration="3m8.383587161s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.375735017 +0000 UTC m=+247.275522716" watchObservedRunningTime="2026-03-13 14:09:08.383587161 +0000 UTC m=+247.283374860" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.422261 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-544df" podStartSLOduration=187.422237741 podStartE2EDuration="3m7.422237741s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.414823 +0000 UTC m=+247.314610689" watchObservedRunningTime="2026-03-13 14:09:08.422237741 +0000 UTC m=+247.322025430" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.431151 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.433215 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.447081 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.448207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.451201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.451445 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.451684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.452930 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:08.952913835 +0000 UTC m=+247.852701524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.453151 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.453292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.458534 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.463904 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.469421 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.469485 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.528005 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"community-operators-v6msv\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.548555 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-f8mmz" podStartSLOduration=187.548537655 podStartE2EDuration="3m7.548537655s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.54689701 +0000 UTC m=+247.446684709" watchObservedRunningTime="2026-03-13 14:09:08.548537655 +0000 UTC m=+247.448325344" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.553284 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.553788 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.553927 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.053905521 +0000 UTC m=+247.953693210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.554384 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.554523 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.554703 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.555143 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.055131274 +0000 UTC m=+247.954918963 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.572204 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" podStartSLOduration=187.572182867 podStartE2EDuration="3m7.572182867s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:08.57154052 +0000 UTC m=+247.471328209" watchObservedRunningTime="2026-03-13 14:09:08.572182867 +0000 UTC m=+247.471970556" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.591240 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655307 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655591 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655662 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.655693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.656324 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.156299445 +0000 UTC m=+248.056087134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.658838 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.659913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.710714 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.712179 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.712515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"certified-operators-42kpk\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.737385 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757429 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757492 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757530 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.757557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.757867 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.257854556 +0000 UTC m=+248.157642245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.770541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.841302 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.852223 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.859870 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.860861 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.860950 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.861022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.861057 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.861802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.867001 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.867105 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.367051554 +0000 UTC m=+248.266839243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.870548 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.871214 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.885209 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f797b1bb-e769-49c4-84ce-93a6fad2df93-metrics-certs\") pod \"network-metrics-daemon-tzmvc\" (UID: \"f797b1bb-e769-49c4-84ce-93a6fad2df93\") " pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.896087 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"community-operators-zbll9\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.966141 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.966751 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.966988 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:08 crc kubenswrapper[4907]: I0313 14:09:08.967035 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:08 crc kubenswrapper[4907]: E0313 14:09:08.967844 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.467824994 +0000 UTC m=+248.367612683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.059179 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.062961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069130 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069474 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.069519 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.070021 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.569988542 +0000 UTC m=+248.469776231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.070109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.070330 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.073339 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-tzmvc" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.086800 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.102756 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"certified-operators-gjdp7\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: W0313 14:09:09.157910 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84c2ada5_ceab_4327_802e_9ae459ac814d.slice/crio-2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0 WatchSource:0}: Error finding container 2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0: Status 404 returned error can't find the container with id 2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.171000 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.171553 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.671538592 +0000 UTC m=+248.571326281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.205086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerStarted","Data":"9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.205667 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.215077 4907 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-4h9q7 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.215151 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.216222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" event={"ID":"ebd59055-333a-46c2-ba30-826296414a4c","Type":"ContainerStarted","Data":"96d1942009f1b8ce51bc3ce645a94bc1cff961168b69aaf647356e722a6a87aa"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.222353 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"e31cc3fa7fa433ae9637724792028e91568f4c8e8b4df50891f201ffb1364ddd"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.235209 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" podStartSLOduration=188.235158972 podStartE2EDuration="3m8.235158972s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.233752894 +0000 UTC m=+248.133540583" watchObservedRunningTime="2026-03-13 14:09:09.235158972 +0000 UTC m=+248.134946661" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.257118 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.261047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" event={"ID":"71c234e9-5f3d-4b0c-94d4-80b4e33a6dda","Type":"ContainerStarted","Data":"bd285704b82eac1d7e6f05e8c5ba3360170af176524b5e7ebe3a2fe98c0557fb"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.272514 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.272846 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.772825726 +0000 UTC m=+248.672613415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.281405 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wjrms" podStartSLOduration=189.281382749 podStartE2EDuration="3m9.281382749s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.280453593 +0000 UTC m=+248.180241282" watchObservedRunningTime="2026-03-13 14:09:09.281382749 +0000 UTC m=+248.181170438" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.281900 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-djph7" event={"ID":"3a661298-271c-4c57-94ea-3d69a27d7ad5","Type":"ContainerStarted","Data":"9a8f3e0d2c67c4386ad218f76fa5059a7db512cd7dfd04bfe2063252587ce4c2"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.315655 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.329321 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" event={"ID":"1be912b5-9d1d-48bd-a0ab-7f601f9eb9da","Type":"ContainerStarted","Data":"a96675938619010b9f097dc54780a8799b01e25aee1e330d3f5306f2909342f1"} Mar 13 14:09:09 crc kubenswrapper[4907]: W0313 14:09:09.331375 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8adb3c1c_bacd_4cca_9796_7ca96624e9f6.slice/crio-04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da WatchSource:0}: Error finding container 04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da: Status 404 returned error can't find the container with id 04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.335822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerStarted","Data":"2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.353219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-f8mmz" event={"ID":"c26ccc9f-e682-4b94-9539-cb2607725f71","Type":"ContainerStarted","Data":"787a19aa46c50e9190e7b8ab051f8b3388c5f04f9cff3b41bfd80250e281730a"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.375340 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerStarted","Data":"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.376478 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.377764 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.877749039 +0000 UTC m=+248.777536728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.395363 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" podStartSLOduration=188.395336616 podStartE2EDuration="3m8.395336616s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.375024285 +0000 UTC m=+248.274811974" watchObservedRunningTime="2026-03-13 14:09:09.395336616 +0000 UTC m=+248.295124305" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.430739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-56rqr" event={"ID":"59980734-2cf7-42dc-9f6f-a6de7e1c3665","Type":"ContainerStarted","Data":"0a47e925b05361e00dba14ee9fb8ae5b234f948b23c9d549a757d4788ccadd08"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.470839 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" event={"ID":"0396c426-1d56-44e6-885a-1b97d69ad57a","Type":"ContainerStarted","Data":"b37cfd7cc95bf8691f1a46c915bb50b3bf1a28b831571a3218c322339ced6591"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.472021 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.472090 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.478799 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.480680 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:09.980663796 +0000 UTC m=+248.880451485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.503851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-l9xpb" podStartSLOduration=188.503823896 podStartE2EDuration="3m8.503823896s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.396318573 +0000 UTC m=+248.296106262" watchObservedRunningTime="2026-03-13 14:09:09.503823896 +0000 UTC m=+248.403611585" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.504224 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-rmkfg" podStartSLOduration=188.504217757 podStartE2EDuration="3m8.504217757s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.501628476 +0000 UTC m=+248.401416165" watchObservedRunningTime="2026-03-13 14:09:09.504217757 +0000 UTC m=+248.404005446" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.542208 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8jqgq" event={"ID":"fa265490-c23b-4b05-94bc-78b4a636e065","Type":"ContainerStarted","Data":"d87d12281854a7a4b9fc5df3cb864ebff9b2ac551f657b745dd2c8769835232b"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.542394 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.560995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" event={"ID":"814322e3-6251-4ea3-a86b-7b89ca6ea728","Type":"ContainerStarted","Data":"8e1acd7b778f9d3afc64ccd1ecf628a736572fda7bb48f3f9f0fce5bf77942f6"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.561039 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" event={"ID":"814322e3-6251-4ea3-a86b-7b89ca6ea728","Type":"ContainerStarted","Data":"a6498ada9850a819ec9f7e5a616b843cca7a2508a87755d09458256c9b442008"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.578516 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" event={"ID":"699da377-c61b-48f0-92e0-42bd247ee363","Type":"ContainerStarted","Data":"fa1cb4cbad0b730f0bb9a1a91d74909267b6ff61df39cbede52916f05f18ad8e"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.580429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.582793 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.082780692 +0000 UTC m=+248.982568381 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.589912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w76kw" event={"ID":"9b75954c-b75d-4f17-82dc-4c0358fd8d0f","Type":"ContainerStarted","Data":"55c6769fb46a55fb50d7205a8ebc8698e80e046f65ffce5e932d9ab6bc93eae2"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.590303 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.619020 4907 patch_prober.go:28] interesting pod/console-operator-58897d9998-w76kw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.619074 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podUID="9b75954c-b75d-4f17-82dc-4c0358fd8d0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.633002 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-8jqgq" podStartSLOduration=9.632982837 podStartE2EDuration="9.632982837s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.62277182 +0000 UTC m=+248.522559509" watchObservedRunningTime="2026-03-13 14:09:09.632982837 +0000 UTC m=+248.532770526" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.646228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" event={"ID":"8013cba2-6b3c-4548-af5c-6dd0057da0c7","Type":"ContainerStarted","Data":"24c210dbfab7f5c7d92cdf58d42ef582dc91f8401eda4940a42300f5067b3f1b"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.682777 4907 generic.go:334] "Generic (PLEG): container finished" podID="5a9c889b-5954-4d3a-ae72-12000ee1cc25" containerID="4651f82e390b3a5ac2c55407710eda6955584915c8001c9a1dc4182061851ee7" exitCode=0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.682900 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerDied","Data":"4651f82e390b3a5ac2c55407710eda6955584915c8001c9a1dc4182061851ee7"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.682948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"bcc0a7feb3950443a39009d894a278b8d6140c0e85b685de66a2bb1189e17496"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.685322 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.685587 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.185565197 +0000 UTC m=+249.085352936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.685821 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.689602 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.189594346 +0000 UTC m=+249.089382035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.715831 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vfbsx" podStartSLOduration=188.715810129 podStartE2EDuration="3m8.715810129s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.661921774 +0000 UTC m=+248.561709463" watchObservedRunningTime="2026-03-13 14:09:09.715810129 +0000 UTC m=+248.615597818" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.717480 4907 generic.go:334] "Generic (PLEG): container finished" podID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerID="d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9" exitCode=0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.717557 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerDied","Data":"d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.739002 4907 generic.go:334] "Generic (PLEG): container finished" podID="0149e083-ed5b-4c7f-94e7-4f3f2ff8a557" containerID="40f0b1a25cd0d0a5030ec1dcae203661c52d45d8711eefed49ad300a5495f0ac" exitCode=0 Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.739772 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerDied","Data":"40f0b1a25cd0d0a5030ec1dcae203661c52d45d8711eefed49ad300a5495f0ac"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.743732 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" event={"ID":"74b29e18-7d73-4ef5-9739-04137dd1b191","Type":"ContainerStarted","Data":"ba1f38dc924a82f64021f6ef06f3ceb7e7b06c76b8f07b0d1bffc24eb269af27"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.781607 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" event={"ID":"ae320273-06e8-43c5-a64f-acd80ad16d1c","Type":"ContainerStarted","Data":"e7cd020657b593a0fe31337cf7b9bcad7216063124dfc53dafa29723e59f8c19"} Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.784045 4907 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5stnc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.784101 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785363 4907 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kwfhg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785394 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785439 4907 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ng4kz container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785451 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785733 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785749 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.785791 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hwf6r" podStartSLOduration=188.785782031 podStartE2EDuration="3m8.785782031s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.718323038 +0000 UTC m=+248.618110737" watchObservedRunningTime="2026-03-13 14:09:09.785782031 +0000 UTC m=+248.685569720" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.787182 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-tzmvc"] Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.787307 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.787510 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.287489738 +0000 UTC m=+249.187277427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.787602 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.787846 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.287840428 +0000 UTC m=+249.187628117 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.816226 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.816258 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.840992 4907 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-ksgr6 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.841041 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" podUID="1be912b5-9d1d-48bd-a0ab-7f601f9eb9da" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.875725 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podStartSLOduration=188.875702666 podStartE2EDuration="3m8.875702666s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.809870116 +0000 UTC m=+248.709657805" watchObservedRunningTime="2026-03-13 14:09:09.875702666 +0000 UTC m=+248.775490355" Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.890153 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:09 crc kubenswrapper[4907]: E0313 14:09:09.905866 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.405819255 +0000 UTC m=+249.305606944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:09 crc kubenswrapper[4907]: I0313 14:09:09.922422 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-92gs7" podStartSLOduration=188.922397646 podStartE2EDuration="3m8.922397646s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.922045877 +0000 UTC m=+248.821833566" watchObservedRunningTime="2026-03-13 14:09:09.922397646 +0000 UTC m=+248.822185335" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.002488 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.003176 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.503155931 +0000 UTC m=+249.402943620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.011621 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ksfb8" podStartSLOduration=189.01157894 podStartE2EDuration="3m9.01157894s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:09.984546605 +0000 UTC m=+248.884334294" watchObservedRunningTime="2026-03-13 14:09:10.01157894 +0000 UTC m=+248.911366629" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.012868 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-5szhc" podStartSLOduration=189.012863375 podStartE2EDuration="3m9.012863375s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:10.003925222 +0000 UTC m=+248.903712911" watchObservedRunningTime="2026-03-13 14:09:10.012863375 +0000 UTC m=+248.912651064" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.107682 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.108481 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.608459584 +0000 UTC m=+249.508247283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.172194 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.185219 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.209268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.209595 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.709584203 +0000 UTC m=+249.609371892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: W0313 14:09:10.237143 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ef16085_0f4d_435d_a559_20a33c237531.slice/crio-5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a WatchSource:0}: Error finding container 5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a: Status 404 returned error can't find the container with id 5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.310951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.311312 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.811295009 +0000 UTC m=+249.711082698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.311423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.311752 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.811743281 +0000 UTC m=+249.711530970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.412252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.412744 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:10.912725346 +0000 UTC m=+249.812513035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.414785 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55010: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.429219 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.430501 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.433774 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.447618 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.478095 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:10 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:10 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:10 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.478618 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.508137 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55018: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513587 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513684 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513738 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.513765 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.514084 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.014072211 +0000 UTC m=+249.913859900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615117 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615312 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55024: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.615331 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.115298194 +0000 UTC m=+250.015085883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615442 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615620 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.615678 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.615718 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.115705885 +0000 UTC m=+250.015493574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.616259 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.616281 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.643159 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"redhat-marketplace-jk5nw\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.701554 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55040: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.717300 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.717495 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.217446191 +0000 UTC m=+250.117233880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.717673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.718051 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.218041307 +0000 UTC m=+250.117828996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.772588 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.814675 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55050: no serving certificate available for the kubelet" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.817736 4907 generic.go:334] "Generic (PLEG): container finished" podID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerID="0bd7973d57fde27ab0e7adf532d86c90333b4abd96119d9c054a7fb89cbb6dd6" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.818370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"0bd7973d57fde27ab0e7adf532d86c90333b4abd96119d9c054a7fb89cbb6dd6"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.818680 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.818975 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.31895735 +0000 UTC m=+250.218745039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.829318 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" event={"ID":"5a9c889b-5954-4d3a-ae72-12000ee1cc25","Type":"ContainerStarted","Data":"fbd6a73b2a9ffd348f53b7ffb74a46e3e4d4a0debddae3fc1d71bbfc0b464b7b"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.837144 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.841839 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.846234 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" event={"ID":"ebd59055-333a-46c2-ba30-826296414a4c","Type":"ContainerStarted","Data":"008bbcbd9d87f286dcc70fceddeacf79d404fa4a3cc2853e4736cb3009c44561"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.864522 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.864593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.864619 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerStarted","Data":"595859a9bf32c878bf56920b076fb65741978b6d9f83d6c4a352d65ef0a38504"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.872889 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.884553 4907 generic.go:334] "Generic (PLEG): container finished" podID="5ef16085-0f4d-435d-a559-20a33c237531" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.884644 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.884672 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerStarted","Data":"5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.899664 4907 generic.go:334] "Generic (PLEG): container finished" podID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerID="0d444f1a14d20e1e6a2efc71bfbe0eb77305492df11ed2559d59993943816ad1" exitCode=0 Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.899750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"0d444f1a14d20e1e6a2efc71bfbe0eb77305492df11ed2559d59993943816ad1"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.899783 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerStarted","Data":"04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.901922 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" event={"ID":"f797b1bb-e769-49c4-84ce-93a6fad2df93","Type":"ContainerStarted","Data":"559726274eea9d689eaef1bf7a9391dd2a474c1381281965a68d454600e7e5c0"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.901959 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" event={"ID":"f797b1bb-e769-49c4-84ce-93a6fad2df93","Type":"ContainerStarted","Data":"1a6c62d66ca1eb5eb6d74358b2a1c9d9d864fc4e20db8677f3c3e89606814482"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.913851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" podStartSLOduration=190.91383832 podStartE2EDuration="3m10.91383832s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:10.912277427 +0000 UTC m=+249.812065116" watchObservedRunningTime="2026-03-13 14:09:10.91383832 +0000 UTC m=+249.813626009" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920374 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920535 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.920592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.922868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" event={"ID":"0149e083-ed5b-4c7f-94e7-4f3f2ff8a557","Type":"ContainerStarted","Data":"9858b9ab2ce7880d483ef8f05b8275700629d4165e882f596ef9ef70ad0538c9"} Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.923983 4907 patch_prober.go:28] interesting pod/console-operator-58897d9998-w76kw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.924048 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podUID="9b75954c-b75d-4f17-82dc-4c0358fd8d0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 13 14:09:10 crc kubenswrapper[4907]: E0313 14:09:10.925049 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.425023944 +0000 UTC m=+250.324811633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.928125 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.929126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.929562 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:10 crc kubenswrapper[4907]: I0313 14:09:10.935897 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55056: no serving certificate available for the kubelet" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.050061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.050443 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.050724 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.051264 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.057289 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.557244568 +0000 UTC m=+250.457032257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.058003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.060811 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.096956 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"redhat-marketplace-p257z\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.123268 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xkdjh" podStartSLOduration=190.123250052 podStartE2EDuration="3m10.123250052s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:11.120388465 +0000 UTC m=+250.020176164" watchObservedRunningTime="2026-03-13 14:09:11.123250052 +0000 UTC m=+250.023037741" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.152314 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.153062 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.653050393 +0000 UTC m=+250.552838082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.155181 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55062: no serving certificate available for the kubelet" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.200566 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.254251 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.255012 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.754995335 +0000 UTC m=+250.654783024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.302531 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.335373 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" podStartSLOduration=191.335353549 podStartE2EDuration="3m11.335353549s" podCreationTimestamp="2026-03-13 14:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:11.301098538 +0000 UTC m=+250.200886227" watchObservedRunningTime="2026-03-13 14:09:11.335353549 +0000 UTC m=+250.235141238" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.355849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.356206 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.856193836 +0000 UTC m=+250.755981525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.443228 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.444205 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.458198 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.458659 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:11.958633601 +0000 UTC m=+250.858421290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.458970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.474650 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:11 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:11 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:11 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.474710 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.481991 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.556991 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55068: no serving certificate available for the kubelet" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562203 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562297 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.562355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.562840 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.062825023 +0000 UTC m=+250.962612702 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.617840 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.667933 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.668153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.668217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.668282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.668682 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.168666841 +0000 UTC m=+251.068454530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.669618 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.669848 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.726411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"redhat-operators-fmzmc\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.737343 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.770995 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") pod \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.771064 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") pod \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.771392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") pod \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\" (UID: \"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.771598 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.772068 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.272052042 +0000 UTC m=+251.171839721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.777313 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume" (OuterVolumeSpecName: "config-volume") pod "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" (UID: "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.782507 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" (UID: "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.783925 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d" (OuterVolumeSpecName: "kube-api-access-h9w5d") pod "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" (UID: "19edf37b-a1d3-4c7c-945e-5d0b8bb601bf"). InnerVolumeSpecName "kube-api-access-h9w5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.792330 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.806604 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.876655 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.877158 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9w5d\" (UniqueName: \"kubernetes.io/projected/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-kube-api-access-h9w5d\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.877180 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:11 crc kubenswrapper[4907]: I0313 14:09:11.877192 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:11 crc kubenswrapper[4907]: E0313 14:09:11.877494 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.377472888 +0000 UTC m=+251.277260577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.101797 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.102435 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.602413973 +0000 UTC m=+251.502201672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.105691 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerStarted","Data":"4fd25f47bfaf1894861177f77553de05bdcc490dd081b002f769378e51f3d5fb"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.108449 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.108790 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerName="collect-profiles" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.108813 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerName="collect-profiles" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.108990 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" containerName="collect-profiles" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.109973 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.122843 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.147708 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" event={"ID":"19edf37b-a1d3-4c7c-945e-5d0b8bb601bf","Type":"ContainerDied","Data":"e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.147762 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8a5550c81aa8c4bb6491a89a3041cb9a6e5701c9a00e500620d96a20f452a36" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.147844 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.150804 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-tzmvc" event={"ID":"f797b1bb-e769-49c4-84ce-93a6fad2df93","Type":"ContainerStarted","Data":"04f494188fd5ff2ecf6aeca648a652d5db8d54525a4fd3526e002308da68d7d6"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.153155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerStarted","Data":"f1572a3e7d8d63236fe0e9ac91306d55a1b117717a36d7026cf70c18e50c6571"} Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.205049 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.205594 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.205759 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.206054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.209924 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.709871025 +0000 UTC m=+251.609658934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.241981 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55084: no serving certificate available for the kubelet" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307130 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307246 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: E0313 14:09:12.307607 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:12.807587871 +0000 UTC m=+251.707375560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.307915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.308086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:12 crc kubenswrapper[4907]: I0313 14:09:12.393257 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"redhat-operators-vx5qd\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:14.998741 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:14.999310 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.999289845 +0000 UTC m=+254.899077524 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.014419 4907 patch_prober.go:28] interesting pod/console-operator-58897d9998-w76kw container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.014811 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w76kw" podUID="9b75954c-b75d-4f17-82dc-4c0358fd8d0f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.034706 4907 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wgpdd container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.034830 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" podUID="0149e083-ed5b-4c7f-94e7-4f3f2ff8a557" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.034998 4907 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-wgpdd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.035021 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" podUID="0149e083-ed5b-4c7f-94e7-4f3f2ff8a557" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.036011 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:15 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.036067 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.073196 4907 patch_prober.go:28] interesting pod/apiserver-76f77b778f-r86r4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.073280 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" podUID="5a9c889b-5954-4d3a-ae72-12000ee1cc25" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.9:8443/livez\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.093415 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.093454 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.093482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-hkbqt" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.097610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.100353 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-tzmvc" podStartSLOduration=194.100326813 podStartE2EDuration="3m14.100326813s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:12.173729723 +0000 UTC m=+251.073517412" watchObservedRunningTime="2026-03-13 14:09:15.100326813 +0000 UTC m=+254.000114502" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.101495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.102050 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.60203051 +0000 UTC m=+254.501818199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.107496 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.107564 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.112583 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.112626 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.126467 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.127571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.168680 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.171404 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.179117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.204342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.204491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.204560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.280941 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.780922636 +0000 UTC m=+254.680710325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.283791 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.284024 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.300543 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55094: no serving certificate available for the kubelet" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.309135 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.809086192 +0000 UTC m=+254.708873881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.308671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.332509 4907 patch_prober.go:28] interesting pod/console-f9d7485db-l9xpb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.334193 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-l9xpb" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.362534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.362982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.363384 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.364466 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.864437797 +0000 UTC m=+254.764225486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.419746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.421568 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.466418 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.470485 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:15 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.470528 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.481623 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.485186 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:15.985147979 +0000 UTC m=+254.884935668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.553779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.574214 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.581409 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.581727 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" containerID="cri-o://11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1" gracePeriod=30 Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.583028 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.584115 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.084099178 +0000 UTC m=+254.983886867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: W0313 14:09:15.596451 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec3466c2_4954_459d_88d8_e8c6cbf309e8.slice/crio-33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2 WatchSource:0}: Error finding container 33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2: Status 404 returned error can't find the container with id 33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2 Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.655367 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.655635 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" containerID="cri-o://9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2" gracePeriod=30 Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.684999 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.18497153 +0000 UTC m=+255.084759219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.685040 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.686037 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.686459 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.186449312 +0000 UTC m=+255.086236991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.720639 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.793225 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.793722 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.293690396 +0000 UTC m=+255.193478085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.895415 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:15 crc kubenswrapper[4907]: E0313 14:09:15.896114 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.396101751 +0000 UTC m=+255.295889440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.943486 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.954791 4907 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-ksgr6 container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]log ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]etcd ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]etcd-readiness ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 13 14:09:15 crc kubenswrapper[4907]: [-]informer-sync failed: reason withheld Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/generic-apiserver-start-informers ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/max-in-flight-filter ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-StartUserInformer ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-StartOAuthInformer ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-StartTokenTimeoutUpdater ok Mar 13 14:09:15 crc kubenswrapper[4907]: [+]shutdown ok Mar 13 14:09:15 crc kubenswrapper[4907]: readyz check failed Mar 13 14:09:15 crc kubenswrapper[4907]: I0313 14:09:15.954873 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" podUID="1be912b5-9d1d-48bd-a0ab-7f601f9eb9da" containerName="oauth-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.000451 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.000648 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.500616202 +0000 UTC m=+255.400403911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.001083 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.001435 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.501423314 +0000 UTC m=+255.401210993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.026570 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-w76kw" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.028183 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.101620 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.102614 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.602582345 +0000 UTC m=+255.502370034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.124380 4907 generic.go:334] "Generic (PLEG): container finished" podID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerID="9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.124468 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerDied","Data":"9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.135192 4907 generic.go:334] "Generic (PLEG): container finished" podID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerID="11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.135307 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerDied","Data":"11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.138941 4907 generic.go:334] "Generic (PLEG): container finished" podID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerID="2f158ff96c417d8c3693d515dfe179799bf07e75ac8f792f12324fd7a18b5c50" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.138995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"2f158ff96c417d8c3693d515dfe179799bf07e75ac8f792f12324fd7a18b5c50"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.145461 4907 generic.go:334] "Generic (PLEG): container finished" podID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerID="167f39bf875dea575d6960ea51e3427665c28d2ee610e574bf41709836edc603" exitCode=0 Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.145729 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"167f39bf875dea575d6960ea51e3427665c28d2ee610e574bf41709836edc603"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.165761 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.165834 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2"} Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.204320 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.204727 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.704707831 +0000 UTC m=+255.604495520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.263432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.295402 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.319837 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.320082 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.820034416 +0000 UTC m=+255.719822115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.321024 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.336405 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.836375771 +0000 UTC m=+255.736163460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.338722 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.351577 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.351939 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.352647 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" containerName="controller-manager" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.353799 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.357176 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.360421 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.362673 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422232 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422396 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422474 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") pod \"814c6dec-6e96-4068-9aec-9e57fefca5db\" (UID: \"814c6dec-6e96-4068-9aec-9e57fefca5db\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.422580 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.422856 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:16.922840291 +0000 UTC m=+255.822627980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.423900 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config" (OuterVolumeSpecName: "config") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.424428 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.424783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca" (OuterVolumeSpecName: "client-ca") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.431409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.434738 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt" (OuterVolumeSpecName: "kube-api-access-gn5jt") pod "814c6dec-6e96-4068-9aec-9e57fefca5db" (UID: "814c6dec-6e96-4068-9aec-9e57fefca5db"). InnerVolumeSpecName "kube-api-access-gn5jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.465164 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:16 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:16 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:16 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.465247 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524350 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524447 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524589 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524601 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524610 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/814c6dec-6e96-4068-9aec-9e57fefca5db-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524619 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn5jt\" (UniqueName: \"kubernetes.io/projected/814c6dec-6e96-4068-9aec-9e57fefca5db-kube-api-access-gn5jt\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.524631 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/814c6dec-6e96-4068-9aec-9e57fefca5db-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.525285 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.025266846 +0000 UTC m=+255.925054535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.625636 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.626135 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.126097767 +0000 UTC m=+256.025885446 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.626709 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.626801 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.126788696 +0000 UTC m=+256.026576385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.642777 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.653423 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.700767 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729576 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729852 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.729978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.730073 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") pod \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\" (UID: \"9bb1952d-0e55-4e45-81c0-69a8cddb827e\") " Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.731564 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca" (OuterVolumeSpecName: "client-ca") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.741236 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config" (OuterVolumeSpecName: "config") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.741354 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.24133627 +0000 UTC m=+256.141123949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.745048 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7" (OuterVolumeSpecName: "kube-api-access-gqtv7") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "kube-api-access-gqtv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.755438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9bb1952d-0e55-4e45-81c0-69a8cddb827e" (UID: "9bb1952d-0e55-4e45-81c0-69a8cddb827e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835642 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835914 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqtv7\" (UniqueName: \"kubernetes.io/projected/9bb1952d-0e55-4e45-81c0-69a8cddb827e-kube-api-access-gqtv7\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835929 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835940 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb1952d-0e55-4e45-81c0-69a8cddb827e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.835950 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9bb1952d-0e55-4e45-81c0-69a8cddb827e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.836199 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.336180279 +0000 UTC m=+256.235967968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.937340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.937483 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.437454362 +0000 UTC m=+256.337242051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:16 crc kubenswrapper[4907]: I0313 14:09:16.937702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:16 crc kubenswrapper[4907]: E0313 14:09:16.938066 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.438058518 +0000 UTC m=+256.337846197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.017704 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.018056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.018093 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.018203 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" containerName="route-controller-manager" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.018581 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.027072 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.027769 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.030425 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.045783 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.048599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.059766 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.559727026 +0000 UTC m=+256.459514715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.060709 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.061321 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.561307178 +0000 UTC m=+256.461095037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.161793 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.162198 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.66213719 +0000 UTC m=+256.561924879 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162443 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162464 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162531 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.162866 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.662846959 +0000 UTC m=+256.562634648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.162994 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.163024 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.163159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.189747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"8baf18c6d4683811fc8b3cb2c3dfe2ac02213d10d0c7cc70876c15379d595455"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.191929 4907 generic.go:334] "Generic (PLEG): container finished" podID="8779a406-c08e-40db-b25d-8c45e784e94d" containerID="8baf18c6d4683811fc8b3cb2c3dfe2ac02213d10d0c7cc70876c15379d595455" exitCode=0 Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.192064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerStarted","Data":"23bfb83692e34549e2b59eb14b5903640d0e07ca09276303909a34702eedee74"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.214991 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.215149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7" event={"ID":"9bb1952d-0e55-4e45-81c0-69a8cddb827e","Type":"ContainerDied","Data":"9aafc231113ab96a27a77794d70a165636d29d17e2ee245224f530f5fe6fe82a"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.215212 4907 scope.go:117] "RemoveContainer" containerID="9435fd54ef37b6b3b6843aafb8ac4987dfdaac5f9146401bd984cd364cd498e2" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.223412 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"909a1142-83a3-43c2-9032-45f79a411200","Type":"ContainerStarted","Data":"01d238bc9f78f8ddbe9c3e9ff416221f6b3b235da4624c67b2cbebfb3bfea1b2"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.228734 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" event={"ID":"814c6dec-6e96-4068-9aec-9e57fefca5db","Type":"ContainerDied","Data":"0f1cf03420439a212080278b489e741330821d26fb8f1062a79437e4aeb8e4c5"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.228838 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kwfhg" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.257665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"597e9fba55f42bf26e073ef22245fee0726c592fd0c8e15ca812b30dbf8bba2e"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.264402 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.264910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.264959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265006 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265054 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.265256 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.265808 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.765776147 +0000 UTC m=+256.665563836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.267725 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.267613 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.268230 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerID="5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40" exitCode=0 Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.268294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40"} Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.268421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.269856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.270441 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.274737 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.277023 4907 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.278813 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.286435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"route-controller-manager-5f8c8c7848-sbqs7\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.287112 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"controller-manager-9c7f75cf9-5jvdt\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.337927 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.348952 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.355567 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-4h9q7"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.363948 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.368618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.368816 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kwfhg"] Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.369795 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.869773985 +0000 UTC m=+256.769561674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.382464 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wgpdd" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.389686 4907 scope.go:117] "RemoveContainer" containerID="11d51b3d36d9283f41d21738ca7a244487cc24be4453cd47b19366b6bff8bcf1" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.405960 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.435850 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.469678 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.470182 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:17.970159604 +0000 UTC m=+256.869947293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.472980 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:17 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:17 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:17 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.473032 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.571189 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.572899 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.072867147 +0000 UTC m=+256.972654836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.678557 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.678791 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.178750895 +0000 UTC m=+257.078538584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.679077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.679470 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.179455734 +0000 UTC m=+257.079243413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.784113 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.784473 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.284414748 +0000 UTC m=+257.184202437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.784604 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.785080 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.285064475 +0000 UTC m=+257.184852164 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.835501 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="814c6dec-6e96-4068-9aec-9e57fefca5db" path="/var/lib/kubelet/pods/814c6dec-6e96-4068-9aec-9e57fefca5db/volumes" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.836852 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb1952d-0e55-4e45-81c0-69a8cddb827e" path="/var/lib/kubelet/pods/9bb1952d-0e55-4e45-81c0-69a8cddb827e/volumes" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.837326 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:17 crc kubenswrapper[4907]: W0313 14:09:17.848577 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3ff9578_cc43_43df_a9f7_7480be932f63.slice/crio-65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536 WatchSource:0}: Error finding container 65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536: Status 404 returned error can't find the container with id 65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536 Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.886720 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.886937 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.386896924 +0000 UTC m=+257.286684633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.887343 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.887869 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.38786061 +0000 UTC m=+257.287648299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.889048 4907 ???:1] "http: TLS handshake error from 192.168.126.11:55096: no serving certificate available for the kubelet" Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.989415 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.989536 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.489513593 +0000 UTC m=+257.389301282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:17 crc kubenswrapper[4907]: I0313 14:09:17.990383 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:17 crc kubenswrapper[4907]: E0313 14:09:17.990953 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 14:09:18.490928832 +0000 UTC m=+257.390716721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-q9z8k" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.020271 4907 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-13T14:09:17.277046235Z","Handler":null,"Name":""} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041579 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041640 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041652 4907 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.041718 4907 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.091185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.097704 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.099529 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 14:09:18 crc kubenswrapper[4907]: W0313 14:09:18.105108 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb17406bc_b62f_4633_a910_cbb294546ed8.slice/crio-f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df WatchSource:0}: Error finding container f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df: Status 404 returned error can't find the container with id f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.192721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.195645 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.195689 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.245199 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-q9z8k\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.297327 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.306123 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.318915 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerStarted","Data":"f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.351017 4907 generic.go:334] "Generic (PLEG): container finished" podID="909a1142-83a3-43c2-9032-45f79a411200" containerID="ae0944c8bd029d03d163fe172529c94d89626bc0ff58ec09f1ac2cc39d977fa6" exitCode=0 Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.352012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"909a1142-83a3-43c2-9032-45f79a411200","Type":"ContainerDied","Data":"ae0944c8bd029d03d163fe172529c94d89626bc0ff58ec09f1ac2cc39d977fa6"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.386426 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"98989dbdfebe1005d6e35af2ea045605979542bf34f85e4631fcd7ebd963e87c"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.395046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerStarted","Data":"c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.395144 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.395168 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerStarted","Data":"65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.399114 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerStarted","Data":"1e6d9369dcd3ed9950da83d6e5af4c20e4ff979f893db5aee4943245de2541e3"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.399149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerStarted","Data":"e609619b64fb7b2f34b3160b59695ed7992caf3eb610ed3befd8b460e3b1a894"} Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.401852 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.417349 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" podStartSLOduration=3.4173127340000002 podStartE2EDuration="3.417312734s" podCreationTimestamp="2026-03-13 14:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:18.415545387 +0000 UTC m=+257.315333076" watchObservedRunningTime="2026-03-13 14:09:18.417312734 +0000 UTC m=+257.317100423" Mar 13 14:09:18 crc kubenswrapper[4907]: I0313 14:09:18.463993 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.463965342 podStartE2EDuration="2.463965342s" podCreationTimestamp="2026-03-13 14:09:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:18.461196627 +0000 UTC m=+257.360984316" watchObservedRunningTime="2026-03-13 14:09:18.463965342 +0000 UTC m=+257.363753031" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.503047 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.526747 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:19 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.526817 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.531843 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:19 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.546325 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.547480 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-8jqgq" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.726100 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" event={"ID":"4a8597a4-c72f-4023-b2fb-1b1237411a36","Type":"ContainerStarted","Data":"c5bf646885b1b8dde36b6dc97c65514d19d56b34f146bae17445453273a1f619"} Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.785605 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-sjwvn" podStartSLOduration=19.785579803 podStartE2EDuration="19.785579803s" podCreationTimestamp="2026-03-13 14:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:19.785036538 +0000 UTC m=+258.684824227" watchObservedRunningTime="2026-03-13 14:09:19.785579803 +0000 UTC m=+258.685367502" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.802436 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.825282 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-ksgr6" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.950169 4907 patch_prober.go:28] interesting pod/apiserver-76f77b778f-r86r4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]log ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]etcd ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/generic-apiserver-start-informers ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/max-in-flight-filter ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 13 14:09:19 crc kubenswrapper[4907]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/project.openshift.io-projectcache ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-startinformers ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 13 14:09:19 crc kubenswrapper[4907]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 13 14:09:19 crc kubenswrapper[4907]: livez check failed Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.950238 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" podUID="5a9c889b-5954-4d3a-ae72-12000ee1cc25" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:19 crc kubenswrapper[4907]: I0313 14:09:19.985186 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:09:20 crc kubenswrapper[4907]: W0313 14:09:20.028059 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c5d91bb_753e_4554_b7a1_6c85fe323875.slice/crio-b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7 WatchSource:0}: Error finding container b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7: Status 404 returned error can't find the container with id b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7 Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.462618 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.476104 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:20 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:20 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:20 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.476173 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.589102 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") pod \"909a1142-83a3-43c2-9032-45f79a411200\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.589531 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "909a1142-83a3-43c2-9032-45f79a411200" (UID: "909a1142-83a3-43c2-9032-45f79a411200"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.589685 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") pod \"909a1142-83a3-43c2-9032-45f79a411200\" (UID: \"909a1142-83a3-43c2-9032-45f79a411200\") " Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.590081 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/909a1142-83a3-43c2-9032-45f79a411200-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.605407 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "909a1142-83a3-43c2-9032-45f79a411200" (UID: "909a1142-83a3-43c2-9032-45f79a411200"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.696143 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/909a1142-83a3-43c2-9032-45f79a411200-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.730398 4907 ???:1] "http: TLS handshake error from 192.168.126.11:35328: no serving certificate available for the kubelet" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.848605 4907 generic.go:334] "Generic (PLEG): container finished" podID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerID="1e6d9369dcd3ed9950da83d6e5af4c20e4ff979f893db5aee4943245de2541e3" exitCode=0 Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.848700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerDied","Data":"1e6d9369dcd3ed9950da83d6e5af4c20e4ff979f893db5aee4943245de2541e3"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.866325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerStarted","Data":"b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.883242 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerStarted","Data":"2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.884818 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.901109 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.923638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"909a1142-83a3-43c2-9032-45f79a411200","Type":"ContainerDied","Data":"01d238bc9f78f8ddbe9c3e9ff416221f6b3b235da4624c67b2cbebfb3bfea1b2"} Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.923715 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01d238bc9f78f8ddbe9c3e9ff416221f6b3b235da4624c67b2cbebfb3bfea1b2" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.923813 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 14:09:20 crc kubenswrapper[4907]: I0313 14:09:20.965477 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" podStartSLOduration=5.96545711 podStartE2EDuration="5.96545711s" podCreationTimestamp="2026-03-13 14:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:20.933563653 +0000 UTC m=+259.833351342" watchObservedRunningTime="2026-03-13 14:09:20.96545711 +0000 UTC m=+259.865244799" Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.473766 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:21 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:21 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:21 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.473896 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.972471 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerStarted","Data":"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa"} Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.973229 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:21 crc kubenswrapper[4907]: I0313 14:09:21.996389 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" podStartSLOduration=200.996362097 podStartE2EDuration="3m20.996362097s" podCreationTimestamp="2026-03-13 14:06:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:21.995030431 +0000 UTC m=+260.894818120" watchObservedRunningTime="2026-03-13 14:09:21.996362097 +0000 UTC m=+260.896149786" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.297954 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.354476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") pod \"e0ad06d9-5751-43e2-bf40-b037e61934f5\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.354645 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") pod \"e0ad06d9-5751-43e2-bf40-b037e61934f5\" (UID: \"e0ad06d9-5751-43e2-bf40-b037e61934f5\") " Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.356042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e0ad06d9-5751-43e2-bf40-b037e61934f5" (UID: "e0ad06d9-5751-43e2-bf40-b037e61934f5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.371257 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e0ad06d9-5751-43e2-bf40-b037e61934f5" (UID: "e0ad06d9-5751-43e2-bf40-b037e61934f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.456110 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e0ad06d9-5751-43e2-bf40-b037e61934f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.456149 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e0ad06d9-5751-43e2-bf40-b037e61934f5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.464779 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:22 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:22 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:22 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.464836 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.994945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e0ad06d9-5751-43e2-bf40-b037e61934f5","Type":"ContainerDied","Data":"e609619b64fb7b2f34b3160b59695ed7992caf3eb610ed3befd8b460e3b1a894"} Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.995452 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e609619b64fb7b2f34b3160b59695ed7992caf3eb610ed3befd8b460e3b1a894" Mar 13 14:09:22 crc kubenswrapper[4907]: I0313 14:09:22.995080 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 14:09:23 crc kubenswrapper[4907]: I0313 14:09:23.040124 4907 ???:1] "http: TLS handshake error from 192.168.126.11:35338: no serving certificate available for the kubelet" Mar 13 14:09:23 crc kubenswrapper[4907]: I0313 14:09:23.464182 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:23 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:23 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:23 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:23 crc kubenswrapper[4907]: I0313 14:09:23.464261 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.465556 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:24 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:24 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:24 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.466032 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.924865 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:24 crc kubenswrapper[4907]: I0313 14:09:24.931245 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-r86r4" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.108559 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.108624 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.108559 4907 patch_prober.go:28] interesting pod/downloads-7954f5f757-544df container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.109700 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-544df" podUID="f8ea4f24-05b0-4661-a617-a4208f9a9188" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.273171 4907 patch_prober.go:28] interesting pod/console-f9d7485db-l9xpb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.273255 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-l9xpb" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.466679 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:25 crc kubenswrapper[4907]: [-]has-synced failed: reason withheld Mar 13 14:09:25 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:25 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:25 crc kubenswrapper[4907]: I0313 14:09:25.466782 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:26 crc kubenswrapper[4907]: I0313 14:09:26.463527 4907 patch_prober.go:28] interesting pod/router-default-5444994796-f8mmz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 14:09:26 crc kubenswrapper[4907]: [+]has-synced ok Mar 13 14:09:26 crc kubenswrapper[4907]: [+]process-running ok Mar 13 14:09:26 crc kubenswrapper[4907]: healthz check failed Mar 13 14:09:26 crc kubenswrapper[4907]: I0313 14:09:26.463603 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-f8mmz" podUID="c26ccc9f-e682-4b94-9539-cb2607725f71" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 14:09:27 crc kubenswrapper[4907]: I0313 14:09:27.464395 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:27 crc kubenswrapper[4907]: I0313 14:09:27.467116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-f8mmz" Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.873329 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.874078 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" containerID="cri-o://c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d" gracePeriod=30 Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.887631 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:31 crc kubenswrapper[4907]: I0313 14:09:31.887903 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" containerID="cri-o://2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73" gracePeriod=30 Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.096066 4907 generic.go:334] "Generic (PLEG): container finished" podID="b17406bc-b62f-4633-a910-cbb294546ed8" containerID="2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73" exitCode=0 Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.096152 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerDied","Data":"2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73"} Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.099271 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerID="c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d" exitCode=0 Mar 13 14:09:33 crc kubenswrapper[4907]: I0313 14:09:33.099328 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerDied","Data":"c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d"} Mar 13 14:09:35 crc kubenswrapper[4907]: I0313 14:09:35.123039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-544df" Mar 13 14:09:35 crc kubenswrapper[4907]: I0313 14:09:35.291199 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:35 crc kubenswrapper[4907]: I0313 14:09:35.296059 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.320102 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.357909 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:37 crc kubenswrapper[4907]: E0313 14:09:37.358219 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358234 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: E0313 14:09:37.358248 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358254 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" Mar 13 14:09:37 crc kubenswrapper[4907]: E0313 14:09:37.358273 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909a1142-83a3-43c2-9032-45f79a411200" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358281 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="909a1142-83a3-43c2-9032-45f79a411200" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358428 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" containerName="route-controller-manager" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358452 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ad06d9-5751-43e2-bf40-b037e61934f5" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.358465 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="909a1142-83a3-43c2-9032-45f79a411200" containerName="pruner" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.359066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.388198 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469066 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469427 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") pod \"b17406bc-b62f-4633-a910-cbb294546ed8\" (UID: \"b17406bc-b62f-4633-a910-cbb294546ed8\") " Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469818 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.469941 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470015 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470022 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config" (OuterVolumeSpecName: "config") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.470385 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.471099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca" (OuterVolumeSpecName: "client-ca") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.479160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.479201 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj" (OuterVolumeSpecName: "kube-api-access-pl7qj") pod "b17406bc-b62f-4633-a910-cbb294546ed8" (UID: "b17406bc-b62f-4633-a910-cbb294546ed8"). InnerVolumeSpecName "kube-api-access-pl7qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572233 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572264 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572311 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b17406bc-b62f-4633-a910-cbb294546ed8-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572324 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b17406bc-b62f-4633-a910-cbb294546ed8-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.572337 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pl7qj\" (UniqueName: \"kubernetes.io/projected/b17406bc-b62f-4633-a910-cbb294546ed8-kube-api-access-pl7qj\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.573610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.574073 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.577466 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.590708 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"route-controller-manager-65cb4586b7-dzz67\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:37 crc kubenswrapper[4907]: I0313 14:09:37.697932 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.143534 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" event={"ID":"b17406bc-b62f-4633-a910-cbb294546ed8","Type":"ContainerDied","Data":"f7b10d1097439391fe4ec620dc14b8023cbc5d2fa40a7a2598511514510245df"} Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.143607 4907 scope.go:117] "RemoveContainer" containerID="2fe15cd6ab7eff3941755a8f44ea7d5a0fd89e8b799f655d40ebc796fb88ee73" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.143651 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.168751 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.172295 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f8c8c7848-sbqs7"] Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.316057 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.408009 4907 patch_prober.go:28] interesting pod/controller-manager-9c7f75cf9-5jvdt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 14:09:38 crc kubenswrapper[4907]: I0313 14:09:38.408231 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:39.789526 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b17406bc-b62f-4633-a910-cbb294546ed8" path="/var/lib/kubelet/pods/b17406bc-b62f-4633-a910-cbb294546ed8/volumes" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.775903 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.805115 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:40 crc kubenswrapper[4907]: E0313 14:09:40.805462 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.805479 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.805628 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" containerName="controller-manager" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.806211 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.829122 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920485 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920561 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") pod \"e3ff9578-cc43-43df-a9f7-7480be932f63\" (UID: \"e3ff9578-cc43-43df-a9f7-7480be932f63\") " Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920798 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.920872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921112 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921166 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921269 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca" (OuterVolumeSpecName: "client-ca") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921637 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.921656 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.922045 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config" (OuterVolumeSpecName: "config") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.940335 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:40 crc kubenswrapper[4907]: I0313 14:09:40.940491 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d" (OuterVolumeSpecName: "kube-api-access-q888d") pod "e3ff9578-cc43-43df-a9f7-7480be932f63" (UID: "e3ff9578-cc43-43df-a9f7-7480be932f63"). InnerVolumeSpecName "kube-api-access-q888d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023197 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023328 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023556 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023612 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3ff9578-cc43-43df-a9f7-7480be932f63-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023623 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q888d\" (UniqueName: \"kubernetes.io/projected/e3ff9578-cc43-43df-a9f7-7480be932f63-kube-api-access-q888d\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.023633 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3ff9578-cc43-43df-a9f7-7480be932f63-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.024452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.024855 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.025772 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.028411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.040603 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"controller-manager-6f9b8d4bb9-c6vxw\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.127831 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.165653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" event={"ID":"e3ff9578-cc43-43df-a9f7-7480be932f63","Type":"ContainerDied","Data":"65e068bf514e5d70b92a415eb6b665b20a27f8eb668a1c226c56693a6080a536"} Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.165755 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt" Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.215783 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.217220 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-9c7f75cf9-5jvdt"] Mar 13 14:09:41 crc kubenswrapper[4907]: I0313 14:09:41.794594 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ff9578-cc43-43df-a9f7-7480be932f63" path="/var/lib/kubelet/pods/e3ff9578-cc43-43df-a9f7-7480be932f63/volumes" Mar 13 14:09:43 crc kubenswrapper[4907]: I0313 14:09:43.542146 4907 ???:1] "http: TLS handshake error from 192.168.126.11:49470: no serving certificate available for the kubelet" Mar 13 14:09:44 crc kubenswrapper[4907]: I0313 14:09:44.606425 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ccjsv" Mar 13 14:09:45 crc kubenswrapper[4907]: E0313 14:09:45.975209 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 14:09:45 crc kubenswrapper[4907]: E0313 14:09:45.975379 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6mhh2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zbll9_openshift-marketplace(4b6ddc05-e51d-4648-84c6-6d1414398a4b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:45 crc kubenswrapper[4907]: E0313 14:09:45.976850 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zbll9" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.145346 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.145539 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:09:46 crc kubenswrapper[4907]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 13 14:09:46 crc kubenswrapper[4907]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9f5cd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29556848-v8kv4_openshift-infra(a8867781-179d-47e4-9a55-330deb10d273): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 13 14:09:46 crc kubenswrapper[4907]: > logger="UnhandledError" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.146769 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" podUID="a8867781-179d-47e4-9a55-330deb10d273" Mar 13 14:09:46 crc kubenswrapper[4907]: E0313 14:09:46.203599 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" podUID="a8867781-179d-47e4-9a55-330deb10d273" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.694115 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zbll9" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.784004 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.784519 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rjgtx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p257z_openshift-marketplace(775414ea-5b94-4ab6-8b7c-fe7c80671ced): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:47 crc kubenswrapper[4907]: E0313 14:09:47.786081 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-p257z" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.041674 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.041755 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.041811 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.042423 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.042499 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b" gracePeriod=600 Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.211610 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b" exitCode=0 Mar 13 14:09:48 crc kubenswrapper[4907]: I0313 14:09:48.211691 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b"} Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.396548 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p257z" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.583645 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.583833 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2x9qs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-42kpk_openshift-marketplace(8adb3c1c-bacd-4cca-9796-7ca96624e9f6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.585091 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-42kpk" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.719222 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.719646 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6q2rc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gjdp7_openshift-marketplace(5ef16085-0f4d-435d-a559-20a33c237531): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:49 crc kubenswrapper[4907]: E0313 14:09:49.720924 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gjdp7" podUID="5ef16085-0f4d-435d-a559-20a33c237531" Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.905175 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.905913 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.906018 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.908504 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 14:09:49 crc kubenswrapper[4907]: I0313 14:09:49.909117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.000760 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.001090 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.102197 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.102250 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.102417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.140711 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:50 crc kubenswrapper[4907]: I0313 14:09:50.226354 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:51 crc kubenswrapper[4907]: I0313 14:09:51.814663 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:51 crc kubenswrapper[4907]: I0313 14:09:51.908875 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.646417 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gjdp7" podUID="5ef16085-0f4d-435d-a559-20a33c237531" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.646829 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-42kpk" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" Mar 13 14:09:53 crc kubenswrapper[4907]: I0313 14:09:53.671061 4907 scope.go:117] "RemoveContainer" containerID="c215281d0e491b45da32a8e52ffe10fe7add03de5080c14587c1be4c05f1e89d" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.813385 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.813866 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-48ddb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fmzmc_openshift-marketplace(ec3466c2-4954-459d-88d8-e8c6cbf309e8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.816171 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fmzmc" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869069 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869482 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-msd8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jk5nw_openshift-marketplace(d342a6a2-43dd-4d15-8aff-f60f9a3383bf): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869754 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.869981 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k6s9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-v6msv_openshift-marketplace(84c2ada5-ceab-4327-802e-9ae459ac814d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.871701 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jk5nw" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.871736 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-v6msv" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.898348 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.898587 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mcf55,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vx5qd_openshift-marketplace(8779a406-c08e-40db-b25d-8c45e784e94d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 14:09:53 crc kubenswrapper[4907]: E0313 14:09:53.900482 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.036315 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.093196 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.093963 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.105142 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.154776 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.155068 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.155113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.250109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de"} Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256530 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256630 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256691 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.256722 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.289871 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerStarted","Data":"c886e79340573e19a4b1d551a8638eac839a10aac09e4c6b2631e0a6b5fec143"} Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.292372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"installer-9-crc\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.304393 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.304714 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jk5nw" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.305454 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.305559 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-v6msv" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.306437 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:54 crc kubenswrapper[4907]: E0313 14:09:54.310344 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fmzmc" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.496674 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:09:54 crc kubenswrapper[4907]: I0313 14:09:54.933170 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.295722 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerStarted","Data":"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.296573 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.295853 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" containerID="cri-o://2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" gracePeriod=30 Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.299113 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerStarted","Data":"a68e3ade1d545da2c68a88db1974b5d105f932a3f0f6695c49e0be2ffa9a3a68"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.305505 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.306047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerStarted","Data":"5c0a54d32578b5bee8b628f6a41ee0c4081016228d46c69bfb1cadd7ed5ecaa2"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.306080 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerStarted","Data":"38a1f0c2ada60deba963941149abcc4b4702a13faf071269ca0e030d1e00d6d1"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.311924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerStarted","Data":"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.311996 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerStarted","Data":"d8255210963d6d02f148a7ef4aea4444e4de09330c2b6ebd0f599bf30f80400d"} Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.312029 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.312107 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" containerID="cri-o://9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" gracePeriod=30 Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.320862 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" podStartSLOduration=24.320842382 podStartE2EDuration="24.320842382s" podCreationTimestamp="2026-03-13 14:09:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:55.317435899 +0000 UTC m=+294.217223608" watchObservedRunningTime="2026-03-13 14:09:55.320842382 +0000 UTC m=+294.220630071" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.321683 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.340837 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" podStartSLOduration=24.340812353 podStartE2EDuration="24.340812353s" podCreationTimestamp="2026-03-13 14:09:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:55.334752789 +0000 UTC m=+294.234540478" watchObservedRunningTime="2026-03-13 14:09:55.340812353 +0000 UTC m=+294.240600052" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.347965 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=6.347944216 podStartE2EDuration="6.347944216s" podCreationTimestamp="2026-03-13 14:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:55.346045645 +0000 UTC m=+294.245833334" watchObservedRunningTime="2026-03-13 14:09:55.347944216 +0000 UTC m=+294.247731905" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.627852 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.657853 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:09:55 crc kubenswrapper[4907]: E0313 14:09:55.658134 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.658151 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.658258 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerName="route-controller-manager" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.658762 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.677802 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.678279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.678335 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.678370 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") pod \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\" (UID: \"7b861e8a-0bd0-4b07-aa40-16b6310b19c1\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.681861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.682059 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.682109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.682219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.683124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config" (OuterVolumeSpecName: "config") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.683772 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca" (OuterVolumeSpecName: "client-ca") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.689577 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.689733 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw" (OuterVolumeSpecName: "kube-api-access-fxxvw") pod "7b861e8a-0bd0-4b07-aa40-16b6310b19c1" (UID: "7b861e8a-0bd0-4b07-aa40-16b6310b19c1"). InnerVolumeSpecName "kube-api-access-fxxvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.712196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.724597 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782572 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782627 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782731 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") pod \"12427c08-13c7-4bad-9dc3-815d65a8860f\" (UID: \"12427c08-13c7-4bad-9dc3-815d65a8860f\") " Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.782959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783140 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783364 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783385 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783394 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxxvw\" (UniqueName: \"kubernetes.io/projected/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-kube-api-access-fxxvw\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783406 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7b861e8a-0bd0-4b07-aa40-16b6310b19c1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783462 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783519 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca" (OuterVolumeSpecName: "client-ca") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.783793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.784174 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.784200 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config" (OuterVolumeSpecName: "config") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.787352 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.796418 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.797107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t" (OuterVolumeSpecName: "kube-api-access-mbr8t") pod "12427c08-13c7-4bad-9dc3-815d65a8860f" (UID: "12427c08-13c7-4bad-9dc3-815d65a8860f"). InnerVolumeSpecName "kube-api-access-mbr8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.804195 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"route-controller-manager-6f8c9bf8d5-trl8q\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.884983 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbr8t\" (UniqueName: \"kubernetes.io/projected/12427c08-13c7-4bad-9dc3-815d65a8860f-kube-api-access-mbr8t\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885025 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885037 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12427c08-13c7-4bad-9dc3-815d65a8860f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885046 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.885054 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12427c08-13c7-4bad-9dc3-815d65a8860f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:55 crc kubenswrapper[4907]: I0313 14:09:55.972904 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.321446 4907 generic.go:334] "Generic (PLEG): container finished" podID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerID="5c0a54d32578b5bee8b628f6a41ee0c4081016228d46c69bfb1cadd7ed5ecaa2" exitCode=0 Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.321525 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerDied","Data":"5c0a54d32578b5bee8b628f6a41ee0c4081016228d46c69bfb1cadd7ed5ecaa2"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325030 4907 generic.go:334] "Generic (PLEG): container finished" podID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" exitCode=0 Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerDied","Data":"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325106 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" event={"ID":"12427c08-13c7-4bad-9dc3-815d65a8860f","Type":"ContainerDied","Data":"d8255210963d6d02f148a7ef4aea4444e4de09330c2b6ebd0f599bf30f80400d"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325123 4907 scope.go:117] "RemoveContainer" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.325117 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328025 4907 generic.go:334] "Generic (PLEG): container finished" podID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" exitCode=0 Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328155 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerDied","Data":"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.328233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67" event={"ID":"7b861e8a-0bd0-4b07-aa40-16b6310b19c1","Type":"ContainerDied","Data":"c886e79340573e19a4b1d551a8638eac839a10aac09e4c6b2631e0a6b5fec143"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.331366 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerStarted","Data":"c818da15368a295079d572ed81796486cb0c9f878bd84d437fa33e0b3e3169f6"} Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.357977 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.361571 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6f9b8d4bb9-c6vxw"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.365419 4907 scope.go:117] "RemoveContainer" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" Mar 13 14:09:56 crc kubenswrapper[4907]: E0313 14:09:56.365901 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae\": container with ID starting with 9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae not found: ID does not exist" containerID="9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.365940 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae"} err="failed to get container status \"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae\": rpc error: code = NotFound desc = could not find container \"9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae\": container with ID starting with 9d0c6bc13b0d5110984df2dac333ed4ee030d297ea1490e3744a9ec0ee9366ae not found: ID does not exist" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.365970 4907 scope.go:117] "RemoveContainer" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.382851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.38282815 podStartE2EDuration="2.38282815s" podCreationTimestamp="2026-03-13 14:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:56.374638169 +0000 UTC m=+295.274425868" watchObservedRunningTime="2026-03-13 14:09:56.38282815 +0000 UTC m=+295.282615839" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.386237 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.392773 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65cb4586b7-dzz67"] Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.398960 4907 scope.go:117] "RemoveContainer" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" Mar 13 14:09:56 crc kubenswrapper[4907]: E0313 14:09:56.400165 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc\": container with ID starting with 2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc not found: ID does not exist" containerID="2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.400223 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc"} err="failed to get container status \"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc\": rpc error: code = NotFound desc = could not find container \"2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc\": container with ID starting with 2807d645f250d96f0d23cefa64683a737d2ab00c01dfb33a7e730110d4ac31cc not found: ID does not exist" Mar 13 14:09:56 crc kubenswrapper[4907]: I0313 14:09:56.420123 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:09:56 crc kubenswrapper[4907]: W0313 14:09:56.422855 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fb34c61_0980_40fc_9c6a_48af425786d0.slice/crio-54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b WatchSource:0}: Error finding container 54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b: Status 404 returned error can't find the container with id 54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.342437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerStarted","Data":"6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d"} Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.342484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerStarted","Data":"54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b"} Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.362004 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" podStartSLOduration=6.361957104 podStartE2EDuration="6.361957104s" podCreationTimestamp="2026-03-13 14:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:09:57.356801764 +0000 UTC m=+296.256589453" watchObservedRunningTime="2026-03-13 14:09:57.361957104 +0000 UTC m=+296.261744793" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.584386 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.607411 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") pod \"a87e7070-cc91-43dd-94bd-63075734c3a8\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.607561 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") pod \"a87e7070-cc91-43dd-94bd-63075734c3a8\" (UID: \"a87e7070-cc91-43dd-94bd-63075734c3a8\") " Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.607837 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a87e7070-cc91-43dd-94bd-63075734c3a8" (UID: "a87e7070-cc91-43dd-94bd-63075734c3a8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.625262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a87e7070-cc91-43dd-94bd-63075734c3a8" (UID: "a87e7070-cc91-43dd-94bd-63075734c3a8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.708919 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a87e7070-cc91-43dd-94bd-63075734c3a8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.708962 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a87e7070-cc91-43dd-94bd-63075734c3a8-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.791044 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" path="/var/lib/kubelet/pods/12427c08-13c7-4bad-9dc3-815d65a8860f/volumes" Mar 13 14:09:57 crc kubenswrapper[4907]: I0313 14:09:57.791939 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b861e8a-0bd0-4b07-aa40-16b6310b19c1" path="/var/lib/kubelet/pods/7b861e8a-0bd0-4b07-aa40-16b6310b19c1/volumes" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.351629 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.351975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a87e7070-cc91-43dd-94bd-63075734c3a8","Type":"ContainerDied","Data":"38a1f0c2ada60deba963941149abcc4b4702a13faf071269ca0e030d1e00d6d1"} Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.352004 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a1f0c2ada60deba963941149abcc4b4702a13faf071269ca0e030d1e00d6d1" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.352129 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.357165 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.536617 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:09:58 crc kubenswrapper[4907]: E0313 14:09:58.536968 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerName="pruner" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.536984 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerName="pruner" Mar 13 14:09:58 crc kubenswrapper[4907]: E0313 14:09:58.536999 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537005 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537111 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87e7070-cc91-43dd-94bd-63075734c3a8" containerName="pruner" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537127 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="12427c08-13c7-4bad-9dc3-815d65a8860f" containerName="controller-manager" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.537541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.539720 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.540340 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.540612 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.540958 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.541031 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.541331 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.547340 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.550508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621495 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621682 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.621724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723260 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.723317 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.724376 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.724967 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.724964 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.729761 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.741025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"controller-manager-87cc55c9b-nlr2j\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:58 crc kubenswrapper[4907]: I0313 14:09:58.858973 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:09:59 crc kubenswrapper[4907]: I0313 14:09:59.277443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:09:59 crc kubenswrapper[4907]: W0313 14:09:59.286726 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf341a2f4_0f47_47cb_b5e6_facbf98694b9.slice/crio-7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378 WatchSource:0}: Error finding container 7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378: Status 404 returned error can't find the container with id 7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378 Mar 13 14:09:59 crc kubenswrapper[4907]: I0313 14:09:59.357794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerStarted","Data":"7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378"} Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.141096 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.142298 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.145633 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.146820 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.242612 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"auto-csr-approver-29556850-cl45g\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.343814 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"auto-csr-approver-29556850-cl45g\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.363955 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"auto-csr-approver-29556850-cl45g\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.366080 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" exitCode=0 Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.366169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98"} Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.370638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerStarted","Data":"695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46"} Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.370870 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.378930 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.417328 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" podStartSLOduration=9.417295381 podStartE2EDuration="9.417295381s" podCreationTimestamp="2026-03-13 14:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:10:00.414546617 +0000 UTC m=+299.314334306" watchObservedRunningTime="2026-03-13 14:10:00.417295381 +0000 UTC m=+299.317083080" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.457731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:00 crc kubenswrapper[4907]: I0313 14:10:00.631261 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:10:00 crc kubenswrapper[4907]: W0313 14:10:00.681774 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a1d486_2cf1_4078_8f0f_2dbc1ea97acc.slice/crio-91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4 WatchSource:0}: Error finding container 91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4: Status 404 returned error can't find the container with id 91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4 Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.241533 4907 csr.go:261] certificate signing request csr-lg4mf is approved, waiting to be issued Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.248039 4907 csr.go:257] certificate signing request csr-lg4mf is issued Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.376037 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556850-cl45g" event={"ID":"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc","Type":"ContainerStarted","Data":"91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4"} Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.378206 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerStarted","Data":"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2"} Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.380973 4907 generic.go:334] "Generic (PLEG): container finished" podID="a8867781-179d-47e4-9a55-330deb10d273" containerID="2e545187909a28cc5890020e1139c1d488e394bab920eab3166b317419887a60" exitCode=0 Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.380992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" event={"ID":"a8867781-179d-47e4-9a55-330deb10d273","Type":"ContainerDied","Data":"2e545187909a28cc5890020e1139c1d488e394bab920eab3166b317419887a60"} Mar 13 14:10:01 crc kubenswrapper[4907]: I0313 14:10:01.399091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zbll9" podStartSLOduration=3.493922748 podStartE2EDuration="53.399066386s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.86826136 +0000 UTC m=+249.768049049" lastFinishedPulling="2026-03-13 14:10:00.773404998 +0000 UTC m=+299.673192687" observedRunningTime="2026-03-13 14:10:01.39437653 +0000 UTC m=+300.294164239" watchObservedRunningTime="2026-03-13 14:10:01.399066386 +0000 UTC m=+300.298854075" Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.249427 4907 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-28 09:01:56.64747356 +0000 UTC Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.249807 4907 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6954h51m54.397670318s for next certificate rotation Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.648820 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.772786 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") pod \"a8867781-179d-47e4-9a55-330deb10d273\" (UID: \"a8867781-179d-47e4-9a55-330deb10d273\") " Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.779579 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd" (OuterVolumeSpecName: "kube-api-access-9f5cd") pod "a8867781-179d-47e4-9a55-330deb10d273" (UID: "a8867781-179d-47e4-9a55-330deb10d273"). InnerVolumeSpecName "kube-api-access-9f5cd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:02 crc kubenswrapper[4907]: I0313 14:10:02.874429 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f5cd\" (UniqueName: \"kubernetes.io/projected/a8867781-179d-47e4-9a55-330deb10d273-kube-api-access-9f5cd\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.393720 4907 generic.go:334] "Generic (PLEG): container finished" podID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerID="7118e101427b23d94455b072a1fe8c88f9a426be2d07a4a232a1f55839394ad6" exitCode=0 Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.393805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556850-cl45g" event={"ID":"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc","Type":"ContainerDied","Data":"7118e101427b23d94455b072a1fe8c88f9a426be2d07a4a232a1f55839394ad6"} Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.397024 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" event={"ID":"a8867781-179d-47e4-9a55-330deb10d273","Type":"ContainerDied","Data":"38620171197d496fb764e2284a8169d5eff52a1213907791553d488eaad39d9d"} Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.397078 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38620171197d496fb764e2284a8169d5eff52a1213907791553d488eaad39d9d" Mar 13 14:10:03 crc kubenswrapper[4907]: I0313 14:10:03.397082 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556848-v8kv4" Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.689910 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.799770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") pod \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\" (UID: \"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc\") " Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.808181 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv" (OuterVolumeSpecName: "kube-api-access-d8fjv") pod "10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" (UID: "10a1d486-2cf1-4078-8f0f-2dbc1ea97acc"). InnerVolumeSpecName "kube-api-access-d8fjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:04 crc kubenswrapper[4907]: I0313 14:10:04.901847 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8fjv\" (UniqueName: \"kubernetes.io/projected/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc-kube-api-access-d8fjv\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.419285 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556850-cl45g" Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.419308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556850-cl45g" event={"ID":"10a1d486-2cf1-4078-8f0f-2dbc1ea97acc","Type":"ContainerDied","Data":"91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4"} Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.419360 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91751e260bdf90455e0e2a20034e5c3e5c8d177cfcae4c3d592f866d7cff94d4" Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.421717 4907 generic.go:334] "Generic (PLEG): container finished" podID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerID="8fae8fbf21c4105d1f6b4e7581d3a799c097d7bce76e220e01e3d861c30432de" exitCode=0 Mar 13 14:10:05 crc kubenswrapper[4907]: I0313 14:10:05.421771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"8fae8fbf21c4105d1f6b4e7581d3a799c097d7bce76e220e01e3d861c30432de"} Mar 13 14:10:06 crc kubenswrapper[4907]: I0313 14:10:06.433016 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerStarted","Data":"f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd"} Mar 13 14:10:06 crc kubenswrapper[4907]: I0313 14:10:06.437153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerStarted","Data":"b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61"} Mar 13 14:10:06 crc kubenswrapper[4907]: I0313 14:10:06.440149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945"} Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.447316 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerID="a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945" exitCode=0 Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.447481 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945"} Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.449773 4907 generic.go:334] "Generic (PLEG): container finished" podID="8779a406-c08e-40db-b25d-8c45e784e94d" containerID="f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd" exitCode=0 Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.449877 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd"} Mar 13 14:10:07 crc kubenswrapper[4907]: I0313 14:10:07.525788 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p257z" podStartSLOduration=7.454787329 podStartE2EDuration="57.525768593s" podCreationTimestamp="2026-03-13 14:09:10 +0000 UTC" firstStartedPulling="2026-03-13 14:09:16.143318192 +0000 UTC m=+255.043105881" lastFinishedPulling="2026-03-13 14:10:06.214299456 +0000 UTC m=+305.114087145" observedRunningTime="2026-03-13 14:10:07.520661924 +0000 UTC m=+306.420449613" watchObservedRunningTime="2026-03-13 14:10:07.525768593 +0000 UTC m=+306.425556282" Mar 13 14:10:09 crc kubenswrapper[4907]: I0313 14:10:09.060483 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:09 crc kubenswrapper[4907]: I0313 14:10:09.060590 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:09 crc kubenswrapper[4907]: I0313 14:10:09.485246 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:10 crc kubenswrapper[4907]: I0313 14:10:10.524144 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.200702 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.202057 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.254320 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.483518 4907 generic.go:334] "Generic (PLEG): container finished" podID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerID="059ebb72320a2144f692825d22d2d54a1e276a63cc41c87a83327001d7f8b539" exitCode=0 Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.483555 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"059ebb72320a2144f692825d22d2d54a1e276a63cc41c87a83327001d7f8b539"} Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.487141 4907 generic.go:334] "Generic (PLEG): container finished" podID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerID="701b8cbaf2465addef85f7946e5e3a30ad0261ab612755a656d7086a119dc2d2" exitCode=0 Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.487257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"701b8cbaf2465addef85f7946e5e3a30ad0261ab612755a656d7086a119dc2d2"} Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.540983 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.825313 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.825590 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" containerID="cri-o://695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46" gracePeriod=30 Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.846477 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:10:11 crc kubenswrapper[4907]: I0313 14:10:11.846778 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" containerID="cri-o://6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d" gracePeriod=30 Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.219839 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.498472 4907 generic.go:334] "Generic (PLEG): container finished" podID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerID="6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d" exitCode=0 Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.498566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerDied","Data":"6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d"} Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.503216 4907 generic.go:334] "Generic (PLEG): container finished" podID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerID="695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46" exitCode=0 Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.503300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerDied","Data":"695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46"} Mar 13 14:10:12 crc kubenswrapper[4907]: I0313 14:10:12.504150 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zbll9" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" containerID="cri-o://80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" gracePeriod=2 Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.073323 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.106814 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.107771 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.107840 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.107868 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108010 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.108028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8867781-179d-47e4-9a55-330deb10d273" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108035 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8867781-179d-47e4-9a55-330deb10d273" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108153 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108189 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8867781-179d-47e4-9a55-330deb10d273" containerName="oc" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.108201 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" containerName="route-controller-manager" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.111549 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.114130 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118244 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118373 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.118408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.127369 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219341 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219385 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219718 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219743 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219762 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219796 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") pod \"7fb34c61-0980-40fc-9c6a-48af425786d0\" (UID: \"7fb34c61-0980-40fc-9c6a-48af425786d0\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219948 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.219999 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.220050 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.220071 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.220417 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config" (OuterVolumeSpecName: "config") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.225020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config" (OuterVolumeSpecName: "config") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.225441 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.225462 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5" (OuterVolumeSpecName: "kube-api-access-hhgz5") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "kube-api-access-hhgz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.226784 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.228404 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x" (OuterVolumeSpecName: "kube-api-access-5gj4x") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "kube-api-access-5gj4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.229037 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.229661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.230467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca" (OuterVolumeSpecName: "client-ca") pod "7fb34c61-0980-40fc-9c6a-48af425786d0" (UID: "7fb34c61-0980-40fc-9c6a-48af425786d0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.236662 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.236746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.239753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"route-controller-manager-596c74884f-nxqmn\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.325803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") pod \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\" (UID: \"f341a2f4-0f47-47cb-b5e6-facbf98694b9\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326767 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca" (OuterVolumeSpecName: "client-ca") pod "f341a2f4-0f47-47cb-b5e6-facbf98694b9" (UID: "f341a2f4-0f47-47cb-b5e6-facbf98694b9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326913 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhgz5\" (UniqueName: \"kubernetes.io/projected/7fb34c61-0980-40fc-9c6a-48af425786d0-kube-api-access-hhgz5\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326940 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326951 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7fb34c61-0980-40fc-9c6a-48af425786d0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326961 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326975 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326985 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gj4x\" (UniqueName: \"kubernetes.io/projected/f341a2f4-0f47-47cb-b5e6-facbf98694b9-kube-api-access-5gj4x\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.326995 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f341a2f4-0f47-47cb-b5e6-facbf98694b9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.327004 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7fb34c61-0980-40fc-9c6a-48af425786d0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.389645 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.428748 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f341a2f4-0f47-47cb-b5e6-facbf98694b9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.440165 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512586 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" exitCode=0 Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512668 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zbll9" event={"ID":"4b6ddc05-e51d-4648-84c6-6d1414398a4b","Type":"ContainerDied","Data":"595859a9bf32c878bf56920b076fb65741978b6d9f83d6c4a352d65ef0a38504"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512687 4907 scope.go:117] "RemoveContainer" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.512785 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zbll9" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.515114 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerStarted","Data":"2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.518451 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerStarted","Data":"c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.521743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" event={"ID":"7fb34c61-0980-40fc-9c6a-48af425786d0","Type":"ContainerDied","Data":"54f13a73875be71a32e36bb0c1839a08275e7ece063f4b0690961fb22cc8775b"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.521809 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.527791 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerStarted","Data":"8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.529462 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") pod \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.529912 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") pod \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.529960 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") pod \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\" (UID: \"4b6ddc05-e51d-4648-84c6-6d1414398a4b\") " Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.530686 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities" (OuterVolumeSpecName: "utilities") pod "4b6ddc05-e51d-4648-84c6-6d1414398a4b" (UID: "4b6ddc05-e51d-4648-84c6-6d1414398a4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.530924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" event={"ID":"f341a2f4-0f47-47cb-b5e6-facbf98694b9","Type":"ContainerDied","Data":"7a696f22a4d1fcb996fa84f332f6a6f73364b4c819e37be9b8353e4327bf6378"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.530984 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87cc55c9b-nlr2j" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.533318 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerStarted","Data":"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.534666 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2" (OuterVolumeSpecName: "kube-api-access-6mhh2") pod "4b6ddc05-e51d-4648-84c6-6d1414398a4b" (UID: "4b6ddc05-e51d-4648-84c6-6d1414398a4b"). InnerVolumeSpecName "kube-api-access-6mhh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.538061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerStarted","Data":"779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.548241 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerStarted","Data":"6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f"} Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.599770 4907 scope.go:117] "RemoveContainer" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.607712 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fmzmc" podStartSLOduration=5.633796244 podStartE2EDuration="1m2.607688756s" podCreationTimestamp="2026-03-13 14:09:11 +0000 UTC" firstStartedPulling="2026-03-13 14:09:16.190183366 +0000 UTC m=+255.089971055" lastFinishedPulling="2026-03-13 14:10:13.164075878 +0000 UTC m=+312.063863567" observedRunningTime="2026-03-13 14:10:13.549790377 +0000 UTC m=+312.449578066" watchObservedRunningTime="2026-03-13 14:10:13.607688756 +0000 UTC m=+312.507476445" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.625957 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6msv" podStartSLOduration=3.28072038 podStartE2EDuration="1m5.62592687s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.830719429 +0000 UTC m=+249.730507118" lastFinishedPulling="2026-03-13 14:10:13.175925919 +0000 UTC m=+312.075713608" observedRunningTime="2026-03-13 14:10:13.607408987 +0000 UTC m=+312.507196676" watchObservedRunningTime="2026-03-13 14:10:13.62592687 +0000 UTC m=+312.525714549" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.631929 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mhh2\" (UniqueName: \"kubernetes.io/projected/4b6ddc05-e51d-4648-84c6-6d1414398a4b-kube-api-access-6mhh2\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.631953 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.633055 4907 scope.go:117] "RemoveContainer" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.665678 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b6ddc05-e51d-4648-84c6-6d1414398a4b" (UID: "4b6ddc05-e51d-4648-84c6-6d1414398a4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.668096 4907 scope.go:117] "RemoveContainer" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.668621 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2\": container with ID starting with 80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2 not found: ID does not exist" containerID="80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.668664 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2"} err="failed to get container status \"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2\": rpc error: code = NotFound desc = could not find container \"80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2\": container with ID starting with 80864b09a701f09fdc5f3a4d57a08fa0f1943de0192a3eda8299785bc1d722e2 not found: ID does not exist" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.668691 4907 scope.go:117] "RemoveContainer" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.669067 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98\": container with ID starting with 0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98 not found: ID does not exist" containerID="0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669099 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98"} err="failed to get container status \"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98\": rpc error: code = NotFound desc = could not find container \"0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98\": container with ID starting with 0a883aeefcfd6a2d4540c0a4e0e6bf02cdfb6c47d06b55f636280edfede41b98 not found: ID does not exist" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669135 4907 scope.go:117] "RemoveContainer" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" Mar 13 14:10:13 crc kubenswrapper[4907]: E0313 14:10:13.669490 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7\": container with ID starting with a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7 not found: ID does not exist" containerID="a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669704 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7"} err="failed to get container status \"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7\": rpc error: code = NotFound desc = could not find container \"a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7\": container with ID starting with a3e791567d7fec33aa07ec52ed2c7bf531de23472ee1120fe2b93a41d79ec2e7 not found: ID does not exist" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.669722 4907 scope.go:117] "RemoveContainer" containerID="6b3f156649bd3aebfecf7ea45641ce94397a3dcd4edc7dc31ce59e515d93132d" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.688301 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-42kpk" podStartSLOduration=3.300493233 podStartE2EDuration="1m5.688276728s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.908383601 +0000 UTC m=+249.808171280" lastFinishedPulling="2026-03-13 14:10:13.296167086 +0000 UTC m=+312.195954775" observedRunningTime="2026-03-13 14:10:13.673693654 +0000 UTC m=+312.573481363" watchObservedRunningTime="2026-03-13 14:10:13.688276728 +0000 UTC m=+312.588064417" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.709470 4907 scope.go:117] "RemoveContainer" containerID="695016fec526a7dbd7db71c59cc85a73528a2dc945096ce528d4e2858fa89f46" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.721911 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vx5qd" podStartSLOduration=7.414619672 podStartE2EDuration="1m2.721866518s" podCreationTimestamp="2026-03-13 14:09:11 +0000 UTC" firstStartedPulling="2026-03-13 14:09:17.19376108 +0000 UTC m=+256.093548769" lastFinishedPulling="2026-03-13 14:10:12.501007926 +0000 UTC m=+311.400795615" observedRunningTime="2026-03-13 14:10:13.700758236 +0000 UTC m=+312.600545925" watchObservedRunningTime="2026-03-13 14:10:13.721866518 +0000 UTC m=+312.621654207" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.724001 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.725688 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f8c9bf8d5-trl8q"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.736554 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b6ddc05-e51d-4648-84c6-6d1414398a4b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.744735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.757461 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-87cc55c9b-nlr2j"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.791504 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fb34c61-0980-40fc-9c6a-48af425786d0" path="/var/lib/kubelet/pods/7fb34c61-0980-40fc-9c6a-48af425786d0/volumes" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.792019 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" path="/var/lib/kubelet/pods/f341a2f4-0f47-47cb-b5e6-facbf98694b9/volumes" Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.875536 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.923924 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:10:13 crc kubenswrapper[4907]: I0313 14:10:13.927652 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zbll9"] Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.017685 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.554355 4907 generic.go:334] "Generic (PLEG): container finished" podID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerID="2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb" exitCode=0 Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.554602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.558707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerStarted","Data":"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.558739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerStarted","Data":"fe5db9115bda6db0c26db9bf40f10d6401ffba671bbe917f6eae2f4ff07fa2ca"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.563586 4907 generic.go:334] "Generic (PLEG): container finished" podID="5ef16085-0f4d-435d-a559-20a33c237531" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" exitCode=0 Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.563605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7"} Mar 13 14:10:14 crc kubenswrapper[4907]: I0313 14:10:14.563835 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p257z" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" containerID="cri-o://b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61" gracePeriod=2 Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.422785 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.422837 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554001 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554229 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-content" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554240 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-content" Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554253 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554259 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554273 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554279 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" Mar 13 14:10:15 crc kubenswrapper[4907]: E0313 14:10:15.554287 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-utilities" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554292 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="extract-utilities" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554416 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" containerName="registry-server" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.554425 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f341a2f4-0f47-47cb-b5e6-facbf98694b9" containerName="controller-manager" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.555020 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.557627 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.557772 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.557914 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.558075 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.558115 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.558118 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.566496 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.570821 4907 generic.go:334] "Generic (PLEG): container finished" podID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerID="b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61" exitCode=0 Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.570919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61"} Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.571379 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.576310 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.581059 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.661963 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.662537 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.662992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.663056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.663104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764696 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764753 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764788 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764833 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.764918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.766043 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.766385 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.766449 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.770781 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.781318 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"controller-manager-6797bc5cfb-75vkh\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.793676 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b6ddc05-e51d-4648-84c6-6d1414398a4b" path="/var/lib/kubelet/pods/4b6ddc05-e51d-4648-84c6-6d1414398a4b/volumes" Mar 13 14:10:15 crc kubenswrapper[4907]: I0313 14:10:15.871993 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.079895 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.105373 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" podStartSLOduration=5.105350064 podStartE2EDuration="5.105350064s" podCreationTimestamp="2026-03-13 14:10:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:10:15.625709451 +0000 UTC m=+314.525497140" watchObservedRunningTime="2026-03-13 14:10:16.105350064 +0000 UTC m=+315.005137753" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.125452 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:16 crc kubenswrapper[4907]: W0313 14:10:16.133478 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod072a25a8_cae1_4572_846b_2d2feea5fb36.slice/crio-381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330 WatchSource:0}: Error finding container 381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330: Status 404 returned error can't find the container with id 381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330 Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.273688 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") pod \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.273740 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") pod \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.273764 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") pod \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\" (UID: \"775414ea-5b94-4ab6-8b7c-fe7c80671ced\") " Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.277651 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities" (OuterVolumeSpecName: "utilities") pod "775414ea-5b94-4ab6-8b7c-fe7c80671ced" (UID: "775414ea-5b94-4ab6-8b7c-fe7c80671ced"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.283063 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx" (OuterVolumeSpecName: "kube-api-access-rjgtx") pod "775414ea-5b94-4ab6-8b7c-fe7c80671ced" (UID: "775414ea-5b94-4ab6-8b7c-fe7c80671ced"). InnerVolumeSpecName "kube-api-access-rjgtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.356456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "775414ea-5b94-4ab6-8b7c-fe7c80671ced" (UID: "775414ea-5b94-4ab6-8b7c-fe7c80671ced"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.375259 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.375310 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/775414ea-5b94-4ab6-8b7c-fe7c80671ced-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.375320 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjgtx\" (UniqueName: \"kubernetes.io/projected/775414ea-5b94-4ab6-8b7c-fe7c80671ced-kube-api-access-rjgtx\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.474842 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" probeResult="failure" output=< Mar 13 14:10:16 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:10:16 crc kubenswrapper[4907]: > Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.578447 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerStarted","Data":"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.580636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p257z" event={"ID":"775414ea-5b94-4ab6-8b7c-fe7c80671ced","Type":"ContainerDied","Data":"f1572a3e7d8d63236fe0e9ac91306d55a1b117717a36d7026cf70c18e50c6571"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.580702 4907 scope.go:117] "RemoveContainer" containerID="b5828a5f740df8bf6f7ad67a6f9e5cd7c2cb33d99d741eccf259274586a3ee61" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.580793 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p257z" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.582607 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerStarted","Data":"de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.584700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerStarted","Data":"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.584742 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerStarted","Data":"381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330"} Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.584923 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.592259 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.595072 4907 scope.go:117] "RemoveContainer" containerID="8fae8fbf21c4105d1f6b4e7581d3a799c097d7bce76e220e01e3d861c30432de" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.610641 4907 scope.go:117] "RemoveContainer" containerID="2f158ff96c417d8c3693d515dfe179799bf07e75ac8f792f12324fd7a18b5c50" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.626405 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jk5nw" podStartSLOduration=6.742675659 podStartE2EDuration="1m6.626387639s" podCreationTimestamp="2026-03-13 14:09:10 +0000 UTC" firstStartedPulling="2026-03-13 14:09:16.152527152 +0000 UTC m=+255.052314841" lastFinishedPulling="2026-03-13 14:10:16.036239132 +0000 UTC m=+314.936026821" observedRunningTime="2026-03-13 14:10:16.622119882 +0000 UTC m=+315.521907571" watchObservedRunningTime="2026-03-13 14:10:16.626387639 +0000 UTC m=+315.526175318" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.651784 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" podStartSLOduration=5.651768796 podStartE2EDuration="5.651768796s" podCreationTimestamp="2026-03-13 14:10:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:10:16.650551573 +0000 UTC m=+315.550339262" watchObservedRunningTime="2026-03-13 14:10:16.651768796 +0000 UTC m=+315.551556485" Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.669957 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:10:16 crc kubenswrapper[4907]: I0313 14:10:16.673752 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p257z"] Mar 13 14:10:17 crc kubenswrapper[4907]: I0313 14:10:17.616818 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gjdp7" podStartSLOduration=4.339136938 podStartE2EDuration="1m9.616798898s" podCreationTimestamp="2026-03-13 14:09:08 +0000 UTC" firstStartedPulling="2026-03-13 14:09:10.889097687 +0000 UTC m=+249.788885376" lastFinishedPulling="2026-03-13 14:10:16.166759627 +0000 UTC m=+315.066547336" observedRunningTime="2026-03-13 14:10:17.615603075 +0000 UTC m=+316.515390784" watchObservedRunningTime="2026-03-13 14:10:17.616798898 +0000 UTC m=+316.516586587" Mar 13 14:10:17 crc kubenswrapper[4907]: I0313 14:10:17.788500 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" path="/var/lib/kubelet/pods/775414ea-5b94-4ab6-8b7c-fe7c80671ced/volumes" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.592966 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.593051 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.648443 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.771726 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.771784 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:18 crc kubenswrapper[4907]: I0313 14:10:18.817510 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.317204 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.317576 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.360308 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.662851 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:10:19 crc kubenswrapper[4907]: I0313 14:10:19.685535 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:10:20 crc kubenswrapper[4907]: I0313 14:10:20.773269 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:20 crc kubenswrapper[4907]: I0313 14:10:20.774009 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:20 crc kubenswrapper[4907]: I0313 14:10:20.830802 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.691251 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.793295 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.793800 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:21 crc kubenswrapper[4907]: I0313 14:10:21.842945 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:22 crc kubenswrapper[4907]: I0313 14:10:22.709797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:10:25 crc kubenswrapper[4907]: I0313 14:10:25.502038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:25 crc kubenswrapper[4907]: I0313 14:10:25.552761 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.422170 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.422450 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vx5qd" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" containerID="cri-o://8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f" gracePeriod=2 Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.684875 4907 generic.go:334] "Generic (PLEG): container finished" podID="8779a406-c08e-40db-b25d-8c45e784e94d" containerID="8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f" exitCode=0 Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.684960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f"} Mar 13 14:10:28 crc kubenswrapper[4907]: I0313 14:10:28.909864 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.085657 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") pod \"8779a406-c08e-40db-b25d-8c45e784e94d\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.085730 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") pod \"8779a406-c08e-40db-b25d-8c45e784e94d\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.085813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") pod \"8779a406-c08e-40db-b25d-8c45e784e94d\" (UID: \"8779a406-c08e-40db-b25d-8c45e784e94d\") " Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.087607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities" (OuterVolumeSpecName: "utilities") pod "8779a406-c08e-40db-b25d-8c45e784e94d" (UID: "8779a406-c08e-40db-b25d-8c45e784e94d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.100288 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55" (OuterVolumeSpecName: "kube-api-access-mcf55") pod "8779a406-c08e-40db-b25d-8c45e784e94d" (UID: "8779a406-c08e-40db-b25d-8c45e784e94d"). InnerVolumeSpecName "kube-api-access-mcf55". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.187526 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.187574 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcf55\" (UniqueName: \"kubernetes.io/projected/8779a406-c08e-40db-b25d-8c45e784e94d-kube-api-access-mcf55\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.209046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8779a406-c08e-40db-b25d-8c45e784e94d" (UID: "8779a406-c08e-40db-b25d-8c45e784e94d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.289513 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8779a406-c08e-40db-b25d-8c45e784e94d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.385557 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.698736 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vx5qd" event={"ID":"8779a406-c08e-40db-b25d-8c45e784e94d","Type":"ContainerDied","Data":"23bfb83692e34549e2b59eb14b5903640d0e07ca09276303909a34702eedee74"} Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.698824 4907 scope.go:117] "RemoveContainer" containerID="8f15859e65d668d8b73b7ccd0f9c53101f5488796a1476b1c18893c5ad25807f" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.698873 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vx5qd" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.723039 4907 scope.go:117] "RemoveContainer" containerID="f50322f1c6601b0cca029ad907edd46e5e3069feeecd3b208e757340796cd8dd" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.751735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.757106 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vx5qd"] Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.767116 4907 scope.go:117] "RemoveContainer" containerID="8baf18c6d4683811fc8b3cb2c3dfe2ac02213d10d0c7cc70876c15379d595455" Mar 13 14:10:29 crc kubenswrapper[4907]: I0313 14:10:29.798488 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" path="/var/lib/kubelet/pods/8779a406-c08e-40db-b25d-8c45e784e94d/volumes" Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.883137 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.883838 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" containerID="cri-o://3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" gracePeriod=30 Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.968350 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:31 crc kubenswrapper[4907]: I0313 14:10:31.968653 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" containerID="cri-o://9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" gracePeriod=30 Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.517205 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ng4kz"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.551737 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.556705 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655744 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655797 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655838 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655872 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655909 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.655982 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.656017 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") pod \"072a25a8-cae1-4572-846b-2d2feea5fb36\" (UID: \"072a25a8-cae1-4572-846b-2d2feea5fb36\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.656044 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") pod \"8c0d9c50-993a-4ade-9001-df91b492e893\" (UID: \"8c0d9c50-993a-4ade-9001-df91b492e893\") " Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.657618 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config" (OuterVolumeSpecName: "config") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.657982 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca" (OuterVolumeSpecName: "client-ca") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.658140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config" (OuterVolumeSpecName: "config") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.658219 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.658570 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca" (OuterVolumeSpecName: "client-ca") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.666766 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp" (OuterVolumeSpecName: "kube-api-access-smvjp") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "kube-api-access-smvjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.679211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.679320 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8c0d9c50-993a-4ade-9001-df91b492e893" (UID: "8c0d9c50-993a-4ade-9001-df91b492e893"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.679390 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt" (OuterVolumeSpecName: "kube-api-access-429xt") pod "072a25a8-cae1-4572-846b-2d2feea5fb36" (UID: "072a25a8-cae1-4572-846b-2d2feea5fb36"). InnerVolumeSpecName "kube-api-access-429xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721579 4907 generic.go:334] "Generic (PLEG): container finished" podID="8c0d9c50-993a-4ade-9001-df91b492e893" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" exitCode=0 Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerDied","Data":"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721697 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" event={"ID":"8c0d9c50-993a-4ade-9001-df91b492e893","Type":"ContainerDied","Data":"fe5db9115bda6db0c26db9bf40f10d6401ffba671bbe917f6eae2f4ff07fa2ca"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721721 4907 scope.go:117] "RemoveContainer" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.721818 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.727960 4907 generic.go:334] "Generic (PLEG): container finished" podID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" exitCode=0 Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.727992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerDied","Data":"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.728029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" event={"ID":"072a25a8-cae1-4572-846b-2d2feea5fb36","Type":"ContainerDied","Data":"381bee78bac3d3aef3f83d5ea311981cbe79eee1108c280c12eebed6b19e7330"} Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.728067 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6797bc5cfb-75vkh" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757689 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c0d9c50-993a-4ade-9001-df91b492e893-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757715 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smvjp\" (UniqueName: \"kubernetes.io/projected/8c0d9c50-993a-4ade-9001-df91b492e893-kube-api-access-smvjp\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757725 4907 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757733 4907 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/072a25a8-cae1-4572-846b-2d2feea5fb36-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757744 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-429xt\" (UniqueName: \"kubernetes.io/projected/072a25a8-cae1-4572-846b-2d2feea5fb36-kube-api-access-429xt\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757753 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757761 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c0d9c50-993a-4ade-9001-df91b492e893-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757770 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.757777 4907 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/072a25a8-cae1-4572-846b-2d2feea5fb36-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.760520 4907 scope.go:117] "RemoveContainer" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" Mar 13 14:10:32 crc kubenswrapper[4907]: E0313 14:10:32.762434 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a\": container with ID starting with 9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a not found: ID does not exist" containerID="9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.762476 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a"} err="failed to get container status \"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a\": rpc error: code = NotFound desc = could not find container \"9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a\": container with ID starting with 9f7688b312e9d673807fef17daaf30d9c92b2663246998695a6ef3925309315a not found: ID does not exist" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.762505 4907 scope.go:117] "RemoveContainer" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.784400 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.788136 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6797bc5cfb-75vkh"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.788227 4907 scope.go:117] "RemoveContainer" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" Mar 13 14:10:32 crc kubenswrapper[4907]: E0313 14:10:32.788630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247\": container with ID starting with 3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247 not found: ID does not exist" containerID="3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.788660 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247"} err="failed to get container status \"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247\": rpc error: code = NotFound desc = could not find container \"3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247\": container with ID starting with 3251605f371bf11c4b8c99a286c7e7a5947870c25059d47601917b31fe721247 not found: ID does not exist" Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.792947 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.795829 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596c74884f-nxqmn"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.819782 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7"] Mar 13 14:10:32 crc kubenswrapper[4907]: I0313 14:10:32.820013 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gjdp7" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" containerID="cri-o://c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" gracePeriod=2 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034195 4907 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034755 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034823 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.034971 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.035051 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.035143 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" gracePeriod=15 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036070 4907 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036329 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036347 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036356 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036363 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036373 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036379 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036387 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036393 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-content" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036400 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036405 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036414 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036419 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036427 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="extract-utilities" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036440 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036446 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036455 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036460 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036469 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036475 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036483 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036488 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036496 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036502 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036509 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036514 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036524 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036531 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036542 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036548 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036560 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036566 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036574 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036582 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036669 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036676 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" containerName="route-controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036685 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036694 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036702 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036708 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036714 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="775414ea-5b94-4ab6-8b7c-fe7c80671ced" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036723 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8779a406-c08e-40db-b25d-8c45e784e94d" containerName="registry-server" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036730 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" containerName="controller-manager" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036738 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036744 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.036829 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036835 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036949 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.036959 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.038575 4907 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.039324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.045711 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.082829 4907 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.162933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163332 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163364 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163580 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163684 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.163766 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.164271 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265536 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265569 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265842 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265855 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.265998 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.266058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.333044 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.333847 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.383957 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:33 crc kubenswrapper[4907]: W0313 14:10:33.407762 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b WatchSource:0}: Error finding container a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b: Status 404 returned error can't find the container with id a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.412860 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c6bf05f43c7db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,LastTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.469445 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") pod \"5ef16085-0f4d-435d-a559-20a33c237531\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.469635 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") pod \"5ef16085-0f4d-435d-a559-20a33c237531\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.469682 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") pod \"5ef16085-0f4d-435d-a559-20a33c237531\" (UID: \"5ef16085-0f4d-435d-a559-20a33c237531\") " Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.472055 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities" (OuterVolumeSpecName: "utilities") pod "5ef16085-0f4d-435d-a559-20a33c237531" (UID: "5ef16085-0f4d-435d-a559-20a33c237531"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.475598 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc" (OuterVolumeSpecName: "kube-api-access-6q2rc") pod "5ef16085-0f4d-435d-a559-20a33c237531" (UID: "5ef16085-0f4d-435d-a559-20a33c237531"). InnerVolumeSpecName "kube-api-access-6q2rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.552358 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ef16085-0f4d-435d-a559-20a33c237531" (UID: "5ef16085-0f4d-435d-a559-20a33c237531"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.571281 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.571338 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ef16085-0f4d-435d-a559-20a33c237531-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.571361 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q2rc\" (UniqueName: \"kubernetes.io/projected/5ef16085-0f4d-435d-a559-20a33c237531-kube-api-access-6q2rc\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.750919 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.753681 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755061 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755102 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755125 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755140 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" exitCode=2 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.755254 4907 scope.go:117] "RemoveContainer" containerID="1c26e675ea96bcff580798e62f5549349a76e259a156c458d1d332e4d4641b80" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.759771 4907 generic.go:334] "Generic (PLEG): container finished" podID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerID="c818da15368a295079d572ed81796486cb0c9f878bd84d437fa33e0b3e3169f6" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.759837 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerDied","Data":"c818da15368a295079d572ed81796486cb0c9f878bd84d437fa33e0b3e3169f6"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.761443 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.761926 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778152 4907 generic.go:334] "Generic (PLEG): container finished" podID="5ef16085-0f4d-435d-a559-20a33c237531" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" exitCode=0 Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778330 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gjdp7" event={"ID":"5ef16085-0f4d-435d-a559-20a33c237531","Type":"ContainerDied","Data":"5380ac5cf0f032bd661fec9fad25b22b2b931ad5cc5954ae70fb9c5f8362990a"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.778477 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gjdp7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.782066 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.783222 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.810455 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="072a25a8-cae1-4572-846b-2d2feea5fb36" path="/var/lib/kubelet/pods/072a25a8-cae1-4572-846b-2d2feea5fb36/volumes" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.811666 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c0d9c50-993a-4ade-9001-df91b492e893" path="/var/lib/kubelet/pods/8c0d9c50-993a-4ade-9001-df91b492e893/volumes" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.812982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a1023123e570e5f57c68d511b33b9e2a5b0ce18e77d1c18f79f77bbf5e2f961b"} Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.813266 4907 scope.go:117] "RemoveContainer" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.814049 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.819171 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.837862 4907 scope.go:117] "RemoveContainer" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.862398 4907 scope.go:117] "RemoveContainer" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.881743 4907 scope.go:117] "RemoveContainer" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.882297 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069\": container with ID starting with c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069 not found: ID does not exist" containerID="c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882364 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069"} err="failed to get container status \"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069\": rpc error: code = NotFound desc = could not find container \"c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069\": container with ID starting with c851c33bb27b562ca9f87e1a35a7d7576757ef40b9f02ebed142608872748069 not found: ID does not exist" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882405 4907 scope.go:117] "RemoveContainer" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.882773 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7\": container with ID starting with 031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7 not found: ID does not exist" containerID="031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882808 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7"} err="failed to get container status \"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7\": rpc error: code = NotFound desc = could not find container \"031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7\": container with ID starting with 031c710b0cf90428d9303b494400b216776d335d2ba2176b760375d7548140a7 not found: ID does not exist" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.882835 4907 scope.go:117] "RemoveContainer" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" Mar 13 14:10:33 crc kubenswrapper[4907]: E0313 14:10:33.885052 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c\": container with ID starting with 7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c not found: ID does not exist" containerID="7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c" Mar 13 14:10:33 crc kubenswrapper[4907]: I0313 14:10:33.885091 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c"} err="failed to get container status \"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c\": rpc error: code = NotFound desc = could not find container \"7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c\": container with ID starting with 7e15a77e2ae6c9b87a9062d382cad40c9ca694c03f35d89b5511dd88c400a20c not found: ID does not exist" Mar 13 14:10:34 crc kubenswrapper[4907]: E0313 14:10:34.004759 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c6bf05f43c7db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,LastTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.795083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4"} Mar 13 14:10:34 crc kubenswrapper[4907]: E0313 14:10:34.796247 4907 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.796240 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.797213 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:34 crc kubenswrapper[4907]: I0313 14:10:34.804172 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.191036 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.192061 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.192280 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.222815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") pod \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.222975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") pod \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.223039 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") pod \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\" (UID: \"b66475a4-2e1f-4b18-b565-3cc70bb3720f\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.223404 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock" (OuterVolumeSpecName: "var-lock") pod "b66475a4-2e1f-4b18-b565-3cc70bb3720f" (UID: "b66475a4-2e1f-4b18-b565-3cc70bb3720f"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.224521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b66475a4-2e1f-4b18-b565-3cc70bb3720f" (UID: "b66475a4-2e1f-4b18-b565-3cc70bb3720f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.230310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b66475a4-2e1f-4b18-b565-3cc70bb3720f" (UID: "b66475a4-2e1f-4b18-b565-3cc70bb3720f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.324745 4907 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.324950 4907 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b66475a4-2e1f-4b18-b565-3cc70bb3720f-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.325058 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b66475a4-2e1f-4b18-b565-3cc70bb3720f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.427904 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.428831 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.429577 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.429930 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.430923 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.527520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.527718 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.527780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528301 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528555 4907 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528587 4907 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.528335 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.629636 4907 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.795991 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.819172 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.820608 4907 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" exitCode=0 Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.820734 4907 scope.go:117] "RemoveContainer" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.821456 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.822583 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.823711 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b66475a4-2e1f-4b18-b565-3cc70bb3720f","Type":"ContainerDied","Data":"a68e3ade1d545da2c68a88db1974b5d105f932a3f0f6695c49e0be2ffa9a3a68"} Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.823755 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a68e3ade1d545da2c68a88db1974b5d105f932a3f0f6695c49e0be2ffa9a3a68" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.823727 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.824635 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.826270 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.826520 4907 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.843653 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.844062 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.844583 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.845598 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.846431 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.847028 4907 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.854222 4907 scope.go:117] "RemoveContainer" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.878263 4907 scope.go:117] "RemoveContainer" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.898973 4907 scope.go:117] "RemoveContainer" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.920814 4907 scope.go:117] "RemoveContainer" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.941726 4907 scope.go:117] "RemoveContainer" containerID="60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.979594 4907 scope.go:117] "RemoveContainer" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.980636 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\": container with ID starting with 93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f not found: ID does not exist" containerID="93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.981061 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f"} err="failed to get container status \"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\": rpc error: code = NotFound desc = could not find container \"93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f\": container with ID starting with 93d8c8c79be9d9ffb72de66fc2f3f9821936023071e95875582b45ccb319601f not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.981365 4907 scope.go:117] "RemoveContainer" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.982355 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\": container with ID starting with e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb not found: ID does not exist" containerID="e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.982417 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb"} err="failed to get container status \"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\": rpc error: code = NotFound desc = could not find container \"e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb\": container with ID starting with e3f2b6756078925f8562b6053b3b83798efb72288f5fd18fff680a4ab0d33feb not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.982455 4907 scope.go:117] "RemoveContainer" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.983395 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\": container with ID starting with be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c not found: ID does not exist" containerID="be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983441 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c"} err="failed to get container status \"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\": rpc error: code = NotFound desc = could not find container \"be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c\": container with ID starting with be40a1a33fa482eab363537afaa901127d560afbc1b5b287ef11e66391a5be4c not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983471 4907 scope.go:117] "RemoveContainer" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.983840 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\": container with ID starting with 1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8 not found: ID does not exist" containerID="1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983929 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8"} err="failed to get container status \"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\": rpc error: code = NotFound desc = could not find container \"1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8\": container with ID starting with 1d1b539bbd918fc50428fc18cabcc4c8a43cc96f220b5692d67cf1c8dfbf8ed8 not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.983956 4907 scope.go:117] "RemoveContainer" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.984506 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\": container with ID starting with d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79 not found: ID does not exist" containerID="d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.984541 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79"} err="failed to get container status \"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\": rpc error: code = NotFound desc = could not find container \"d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79\": container with ID starting with d4b503aa3b13268b3688d067f3e456c86afac1fda171aced0990899c3eb89b79 not found: ID does not exist" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.984566 4907 scope.go:117] "RemoveContainer" containerID="60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d" Mar 13 14:10:35 crc kubenswrapper[4907]: E0313 14:10:35.984967 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\": container with ID starting with 60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d not found: ID does not exist" containerID="60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d" Mar 13 14:10:35 crc kubenswrapper[4907]: I0313 14:10:35.985014 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d"} err="failed to get container status \"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\": rpc error: code = NotFound desc = could not find container \"60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d\": container with ID starting with 60ce82af98096367803c824ad0ba4272ade372a31d309c381cf6a452608d610d not found: ID does not exist" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.950321 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.951422 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.951825 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.952318 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.952744 4907 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:40 crc kubenswrapper[4907]: I0313 14:10:40.952785 4907 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 13 14:10:40 crc kubenswrapper[4907]: E0313 14:10:40.953228 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="200ms" Mar 13 14:10:41 crc kubenswrapper[4907]: E0313 14:10:41.154932 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="400ms" Mar 13 14:10:41 crc kubenswrapper[4907]: E0313 14:10:41.556542 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="800ms" Mar 13 14:10:41 crc kubenswrapper[4907]: I0313 14:10:41.787197 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:41 crc kubenswrapper[4907]: I0313 14:10:41.787821 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.357008 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="1.6s" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.780290 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T14:10:42Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.780651 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.781137 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.781398 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.781975 4907 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.782041 4907 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:10:42 crc kubenswrapper[4907]: I0313 14:10:42.940211 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:10:42 crc kubenswrapper[4907]: W0313 14:10:42.940656 4907 reflector.go:561] object-"openshift-network-diagnostics"/"kube-root-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.940713 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:42 crc kubenswrapper[4907]: W0313 14:10:42.940738 4907 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin-cert": failed to list *v1.Secret: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.940870 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:42 crc kubenswrapper[4907]: W0313 14:10:42.941145 4907 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:42 crc kubenswrapper[4907]: E0313 14:10:42.941259 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/configmaps?fieldSelector=metadata.name%3Dnetworking-console-plugin&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.781742 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.783008 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.783418 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.807087 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.807123 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.807360 4907 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.807920 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.863594 4907 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod4b6ddc05-e51d-4648-84c6-6d1414398a4b"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod4b6ddc05-e51d-4648-84c6-6d1414398a4b] : Timed out while waiting for systemd to remove kubepods-burstable-pod4b6ddc05_e51d_4648_84c6_6d1414398a4b.slice" Mar 13 14:10:43 crc kubenswrapper[4907]: I0313 14:10:43.885235 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6b2431eedeb3a3f40d4173ed9069eeef6a78c892216287534c57a19ae6b0eec9"} Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940679 4907 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: failed to sync secret cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.941176 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:45.941147757 +0000 UTC m=+464.840935456 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync secret cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940714 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940730 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.940796 4907 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.941341 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:45.941324112 +0000 UTC m=+464.841111801 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:43 crc kubenswrapper[4907]: W0313 14:10:43.941703 4907 reflector.go:561] object-"openshift-network-diagnostics"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.941797 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:43 crc kubenswrapper[4907]: E0313 14:10:43.957766 4907 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.174:6443: connect: connection refused" interval="3.2s" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.005611 4907 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.174:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c6bf05f43c7db openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,LastTimestamp:2026-03-13 14:10:33.411733467 +0000 UTC m=+332.311521186,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.844673 4907 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" volumeName="registry-storage" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.896281 4907 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="ef2c70913dd6f3338ee16f76a9023591596215ccf7134159f4cb7341cffbf0c1" exitCode=0 Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.896387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"ef2c70913dd6f3338ee16f76a9023591596215ccf7134159f4cb7341cffbf0c1"} Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.897002 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.897056 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.898157 4907 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.898156 4907 status_manager.go:851] "Failed to get status for pod" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:44 crc kubenswrapper[4907]: I0313 14:10:44.899115 4907 status_manager.go:851] "Failed to get status for pod" podUID="5ef16085-0f4d-435d-a559-20a33c237531" pod="openshift-marketplace/certified-operators-gjdp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gjdp7\": dial tcp 38.102.83.174:6443: connect: connection refused" Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942366 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942418 4907 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942445 4907 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942490 4907 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942529 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:46.942494918 +0000 UTC m=+465.842282637 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:44 crc kubenswrapper[4907]: E0313 14:10:44.942591 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 14:12:46.94255852 +0000 UTC m=+465.842346249 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : failed to sync configmap cache: timed out waiting for the condition Mar 13 14:10:45 crc kubenswrapper[4907]: W0313 14:10:45.060966 4907 reflector.go:561] object-"openshift-network-console"/"networking-console-plugin-cert": failed to list *v1.Secret: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253": dial tcp 38.102.83.174:6443: connect: connection refused Mar 13 14:10:45 crc kubenswrapper[4907]: E0313 14:10:45.061140 4907 reflector.go:158] "Unhandled Error" err="object-\"openshift-network-console\"/\"networking-console-plugin-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-console/secrets?fieldSelector=metadata.name%3Dnetworking-console-plugin-cert&resourceVersion=27253\": dial tcp 38.102.83.174:6443: connect: connection refused" logger="UnhandledError" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.340573 4907 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.341110 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.911198 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b74496d1c4984a1e4f4d9b05b79122c2e6c035b15b2f018289763da1eeeeb60c"} Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.911259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"05a5708731085b175d3a5f8b7c9c7da805199496debf2089469cf2e6f29cd42c"} Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.917146 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919074 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919144 4907 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887" exitCode=1 Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887"} Mar 13 14:10:45 crc kubenswrapper[4907]: I0313 14:10:45.919674 4907 scope.go:117] "RemoveContainer" containerID="18eadc065bf4a307b7b443fcfe79b2b719938ee486b53321c0a9f20926a0a887" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.927560 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.929516 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.929620 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ce3d04da2d6730474f09c50e099717f844702dbedb6d0bcf6e1beaf09bfaa48e"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"abae9b949d1bc83d19f3001605ff34418df6c5d957b611db5e92a058405e1b89"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932569 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"065c5409509743840539ff7778a7053b2a294210fadb97a6f1663e20bd78a497"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932581 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"08143c76ef7b653c5bfee9a2778f3a8be16e2709753dcfd74f0f43dc83614561"} Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932816 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.932835 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:46 crc kubenswrapper[4907]: I0313 14:10:46.933045 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:48 crc kubenswrapper[4907]: I0313 14:10:48.809031 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:48 crc kubenswrapper[4907]: I0313 14:10:48.809124 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:48 crc kubenswrapper[4907]: I0313 14:10:48.818814 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:50 crc kubenswrapper[4907]: I0313 14:10:50.075421 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:10:50 crc kubenswrapper[4907]: I0313 14:10:50.083485 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:10:50 crc kubenswrapper[4907]: I0313 14:10:50.961793 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.355387 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.850594 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.851431 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.853769 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.943187 4907 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.968519 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.968569 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:51 crc kubenswrapper[4907]: I0313 14:10:51.973489 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.008128 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="86097500-e6a8-4abd-8b9d-7198222956b2" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.976441 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.977963 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:10:52 crc kubenswrapper[4907]: I0313 14:10:52.983117 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="86097500-e6a8-4abd-8b9d-7198222956b2" Mar 13 14:10:55 crc kubenswrapper[4907]: E0313 14:10:55.811917 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-s2dwl], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 14:10:55 crc kubenswrapper[4907]: E0313 14:10:55.832248 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[networking-console-plugin-cert nginx-conf], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 14:10:55 crc kubenswrapper[4907]: E0313 14:10:55.845709 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-cqllr], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 14:10:57 crc kubenswrapper[4907]: I0313 14:10:57.561272 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" containerID="cri-o://3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc" gracePeriod=15 Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.009807 4907 generic.go:334] "Generic (PLEG): container finished" podID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerID="3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc" exitCode=0 Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.009907 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerDied","Data":"3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc"} Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.101194 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.270865 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.270945 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271182 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271234 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271292 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271326 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271355 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271388 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271481 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271541 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.271578 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") pod \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\" (UID: \"3661f6b2-1d5e-42f0-947f-a32b66258a1f\") " Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.273619 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.274193 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.275965 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.277452 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.279292 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.290020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.291990 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9" (OuterVolumeSpecName: "kube-api-access-mtqz9") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "kube-api-access-mtqz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.292086 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.292384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.292951 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.293568 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.294397 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.298607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.299037 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3661f6b2-1d5e-42f0-947f-a32b66258a1f" (UID: "3661f6b2-1d5e-42f0-947f-a32b66258a1f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379386 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtqz9\" (UniqueName: \"kubernetes.io/projected/3661f6b2-1d5e-42f0-947f-a32b66258a1f-kube-api-access-mtqz9\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379445 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379473 4907 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379498 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379520 4907 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3661f6b2-1d5e-42f0-947f-a32b66258a1f-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379540 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379561 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379583 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379604 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379626 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379647 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379669 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379688 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:58 crc kubenswrapper[4907]: I0313 14:10:58.379709 4907 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3661f6b2-1d5e-42f0-947f-a32b66258a1f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:10:59 crc kubenswrapper[4907]: I0313 14:10:59.020351 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" event={"ID":"3661f6b2-1d5e-42f0-947f-a32b66258a1f","Type":"ContainerDied","Data":"9696c0da9ec098e8a51fc14c33a99177653e98a7c7e1cea2ab3b0156346c1506"} Mar 13 14:10:59 crc kubenswrapper[4907]: I0313 14:10:59.020467 4907 scope.go:117] "RemoveContainer" containerID="3d60ed9e1e0ff620e3a27d58d5fb879793f36dc7e7cd6201ecb90667f5bab7dc" Mar 13 14:10:59 crc kubenswrapper[4907]: I0313 14:10:59.020474 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ng4kz" Mar 13 14:11:01 crc kubenswrapper[4907]: I0313 14:11:01.614299 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.092526 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.590477 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.642844 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 14:11:02 crc kubenswrapper[4907]: I0313 14:11:02.884269 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.257446 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.351096 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.373265 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.549420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 14:11:03 crc kubenswrapper[4907]: I0313 14:11:03.742632 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 14:11:04 crc kubenswrapper[4907]: I0313 14:11:04.687076 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 14:11:04 crc kubenswrapper[4907]: I0313 14:11:04.940681 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.288662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.343630 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.395263 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.729534 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.732996 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.893811 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 14:11:05 crc kubenswrapper[4907]: I0313 14:11:05.957420 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.253602 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.258147 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.303586 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.394396 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.535331 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.636113 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.713698 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.747463 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 14:11:06 crc kubenswrapper[4907]: I0313 14:11:06.860765 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.138016 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.157136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.161128 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.445471 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.521372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.528676 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.537550 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.594021 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.640966 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.676989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.806512 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 14:11:07 crc kubenswrapper[4907]: I0313 14:11:07.910567 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.076807 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.137640 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.245335 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.254180 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.259957 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.303719 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.521811 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.700092 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.781342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:11:08 crc kubenswrapper[4907]: I0313 14:11:08.855603 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.015147 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.039513 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.051768 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.067228 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.071065 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.080125 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.098358 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.098517 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.118130 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.217564 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.315049 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.338061 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.349179 4907 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.351720 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.354406 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gjdp7","openshift-authentication/oauth-openshift-558db77b4-ng4kz","openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.354467 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc","openshift-kube-apiserver/kube-apiserver-crc","openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b","openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp"] Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355003 4907 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355221 4907 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="c5d2b054-0ed8-419c-b731-f14d3598372f" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355447 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerName="installer" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355476 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerName="installer" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355495 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355502 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355521 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-content" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355528 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-content" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355541 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355548 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" Mar 13 14:11:09 crc kubenswrapper[4907]: E0313 14:11:09.355561 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-utilities" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355569 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="extract-utilities" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355738 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66475a4-2e1f-4b18-b565-3cc70bb3720f" containerName="installer" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355749 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" containerName="oauth-openshift" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.355761 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef16085-0f4d-435d-a559-20a33c237531" containerName="registry-server" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.357900 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.358617 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.360092 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.360974 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.361236 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.361684 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365122 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365302 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365378 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365408 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365584 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365742 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365873 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.365993 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366187 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366277 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366320 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366370 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366296 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366397 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366434 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366581 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366684 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366709 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366745 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.366963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.367231 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.367427 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.381271 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.381327 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.383378 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.389414 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.391129 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.443101 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.443083684 podStartE2EDuration="18.443083684s" podCreationTimestamp="2026-03-13 14:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:09.440236688 +0000 UTC m=+368.340024377" watchObservedRunningTime="2026-03-13 14:11:09.443083684 +0000 UTC m=+368.342871373" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.445419 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.528253 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.538352 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.540830 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-config\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b98r7\" (UniqueName: \"kubernetes.io/projected/a704bdfc-46ff-4a8a-8d31-21a140b49c99-kube-api-access-b98r7\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541257 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541481 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541741 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwxs7\" (UniqueName: \"kubernetes.io/projected/03ca839d-b5ba-4b55-87b3-48abca197163-kube-api-access-lwxs7\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541851 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.541988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ca839d-b5ba-4b55-87b3-48abca197163-serving-cert\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542215 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-proxy-ca-bundles\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542320 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542418 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542625 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-config\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-client-ca\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542841 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a704bdfc-46ff-4a8a-8d31-21a140b49c99-serving-cert\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.542960 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-policies\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543107 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-dir\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543240 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543265 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wml9p\" (UniqueName: \"kubernetes.io/projected/8746df2e-524e-42a7-9ca5-c5300c0a1a09-kube-api-access-wml9p\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.543291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-client-ca\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644116 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwxs7\" (UniqueName: \"kubernetes.io/projected/03ca839d-b5ba-4b55-87b3-48abca197163-kube-api-access-lwxs7\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644139 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644172 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ca839d-b5ba-4b55-87b3-48abca197163-serving-cert\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-proxy-ca-bundles\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644229 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644248 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644266 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-config\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-client-ca\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a704bdfc-46ff-4a8a-8d31-21a140b49c99-serving-cert\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644359 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-policies\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-dir\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644427 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wml9p\" (UniqueName: \"kubernetes.io/projected/8746df2e-524e-42a7-9ca5-c5300c0a1a09-kube-api-access-wml9p\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644445 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-client-ca\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644476 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-config\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b98r7\" (UniqueName: \"kubernetes.io/projected/a704bdfc-46ff-4a8a-8d31-21a140b49c99-kube-api-access-b98r7\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644510 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644529 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.644550 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.645374 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-policies\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.645522 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-client-ca\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.645837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-config\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.646480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.646609 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-service-ca\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.647425 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a704bdfc-46ff-4a8a-8d31-21a140b49c99-client-ca\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.647830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8746df2e-524e-42a7-9ca5-c5300c0a1a09-audit-dir\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.647996 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.650026 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-config\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.651208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/03ca839d-b5ba-4b55-87b3-48abca197163-proxy-ca-bundles\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.651649 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03ca839d-b5ba-4b55-87b3-48abca197163-serving-cert\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.652986 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-error\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.652991 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.654530 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.654704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.656660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-login\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.658032 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-router-certs\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.659652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.662113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-session\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.662355 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.663423 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a704bdfc-46ff-4a8a-8d31-21a140b49c99-serving-cert\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.664280 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwxs7\" (UniqueName: \"kubernetes.io/projected/03ca839d-b5ba-4b55-87b3-48abca197163-kube-api-access-lwxs7\") pod \"controller-manager-5f4c4b8dd4-9sl4b\" (UID: \"03ca839d-b5ba-4b55-87b3-48abca197163\") " pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.667997 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wml9p\" (UniqueName: \"kubernetes.io/projected/8746df2e-524e-42a7-9ca5-c5300c0a1a09-kube-api-access-wml9p\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.668086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b98r7\" (UniqueName: \"kubernetes.io/projected/a704bdfc-46ff-4a8a-8d31-21a140b49c99-kube-api-access-b98r7\") pod \"route-controller-manager-5c879c7f8d-2mxxc\" (UID: \"a704bdfc-46ff-4a8a-8d31-21a140b49c99\") " pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.668465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8746df2e-524e-42a7-9ca5-c5300c0a1a09-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7c65b78cb7-ld7fp\" (UID: \"8746df2e-524e-42a7-9ca5-c5300c0a1a09\") " pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.686129 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.692535 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.703238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.708490 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.748907 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.782247 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.790793 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3661f6b2-1d5e-42f0-947f-a32b66258a1f" path="/var/lib/kubelet/pods/3661f6b2-1d5e-42f0-947f-a32b66258a1f/volumes" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.791634 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef16085-0f4d-435d-a559-20a33c237531" path="/var/lib/kubelet/pods/5ef16085-0f4d-435d-a559-20a33c237531/volumes" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.831220 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.867158 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.884037 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 14:11:09 crc kubenswrapper[4907]: I0313 14:11:09.931865 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.034709 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.045492 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.119710 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.216638 4907 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.247723 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.252099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.254390 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.363764 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.381855 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.407209 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.468650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.475305 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.570621 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.672586 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.676418 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.740603 4907 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.781788 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:11:10 crc kubenswrapper[4907]: I0313 14:11:10.948685 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.125603 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.185056 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.291594 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.413974 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.559392 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.584267 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.681723 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.749628 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.759330 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.997266 4907 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:11 crc kubenswrapper[4907]: I0313 14:11:11.997785 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.036063 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.061494 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.196125 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.243853 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.288125 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.353871 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.427532 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.559401 4907 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.587241 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.630182 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.863034 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.919249 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.943467 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:11:12 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621" Netns:"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod "route-controller-manager-5c879c7f8d-2mxxc" not found Mar 13 14:11:12 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:12 crc kubenswrapper[4907]: > Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.943596 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:11:12 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621" Netns:"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod "route-controller-manager-5c879c7f8d-2mxxc" not found Mar 13 14:11:12 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:12 crc kubenswrapper[4907]: > pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.943640 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 14:11:12 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621" Netns:"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99" Path:"" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod "route-controller-manager-5c879c7f8d-2mxxc" not found Mar 13 14:11:12 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:12 crc kubenswrapper[4907]: > pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:12 crc kubenswrapper[4907]: E0313 14:11:12.944540 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager(a704bdfc-46ff-4a8a-8d31-21a140b49c99)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager(a704bdfc-46ff-4a8a-8d31-21a140b49c99)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-5c879c7f8d-2mxxc_openshift-route-controller-manager_a704bdfc-46ff-4a8a-8d31-21a140b49c99_0(f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621): error adding pod openshift-route-controller-manager_route-controller-manager-5c879c7f8d-2mxxc to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621\\\" Netns:\\\"/var/run/netns/208f84ca-664c-4bf4-addd-1a01472d31bb\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-5c879c7f8d-2mxxc;K8S_POD_INFRA_CONTAINER_ID=f75e30dd6fbe4d97da9cd85143325c74ecdd86dbb3207f61697dca55b64c1621;K8S_POD_UID=a704bdfc-46ff-4a8a-8d31-21a140b49c99\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc] networking: Multus: [openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc/a704bdfc-46ff-4a8a-8d31-21a140b49c99]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod route-controller-manager-5c879c7f8d-2mxxc in out of cluster comm: pod \\\"route-controller-manager-5c879c7f8d-2mxxc\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" podUID="a704bdfc-46ff-4a8a-8d31-21a140b49c99" Mar 13 14:11:12 crc kubenswrapper[4907]: I0313 14:11:12.967857 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000229 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e" Netns:"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod "oauth-openshift-7c65b78cb7-ld7fp" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000300 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e" Netns:"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod "oauth-openshift-7c65b78cb7-ld7fp" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000331 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e" Netns:"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod "oauth-openshift-7c65b78cb7-ld7fp" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.000396 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication(8746df2e-524e-42a7-9ca5-c5300c0a1a09)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication(8746df2e-524e-42a7-9ca5-c5300c0a1a09)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7c65b78cb7-ld7fp_openshift-authentication_8746df2e-524e-42a7-9ca5-c5300c0a1a09_0(c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e): error adding pod openshift-authentication_oauth-openshift-7c65b78cb7-ld7fp to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e\\\" Netns:\\\"/var/run/netns/1829481a-098a-4a48-b8c3-231563e389b5\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7c65b78cb7-ld7fp;K8S_POD_INFRA_CONTAINER_ID=c45269bc024ec530d94f1f1c216da78b5947c8922810d8bce846a0c2e15b9b0e;K8S_POD_UID=8746df2e-524e-42a7-9ca5-c5300c0a1a09\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp] networking: Multus: [openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp/8746df2e-524e-42a7-9ca5-c5300c0a1a09]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-7c65b78cb7-ld7fp in out of cluster comm: pod \\\"oauth-openshift-7c65b78cb7-ld7fp\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" podUID="8746df2e-524e-42a7-9ca5-c5300c0a1a09" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.003558 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038713 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df" Netns:"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod "controller-manager-5f4c4b8dd4-9sl4b" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038837 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df" Netns:"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod "controller-manager-5f4c4b8dd4-9sl4b" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038869 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 14:11:13 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df" Netns:"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163" Path:"" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod "controller-manager-5f4c4b8dd4-9sl4b" not found Mar 13 14:11:13 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:11:13 crc kubenswrapper[4907]: > pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:13 crc kubenswrapper[4907]: E0313 14:11:13.038987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager(03ca839d-b5ba-4b55-87b3-48abca197163)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager(03ca839d-b5ba-4b55-87b3-48abca197163)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-5f4c4b8dd4-9sl4b_openshift-controller-manager_03ca839d-b5ba-4b55-87b3-48abca197163_0(5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df): error adding pod openshift-controller-manager_controller-manager-5f4c4b8dd4-9sl4b to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df\\\" Netns:\\\"/var/run/netns/5a26f034-0cf7-47c9-babd-1832d27300bb\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-5f4c4b8dd4-9sl4b;K8S_POD_INFRA_CONTAINER_ID=5c1ba8ff43ed33e226001737ee4130c9659d7444a294f4dcc3db6844b59f44df;K8S_POD_UID=03ca839d-b5ba-4b55-87b3-48abca197163\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b] networking: Multus: [openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b/03ca839d-b5ba-4b55-87b3-48abca197163]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod controller-manager-5f4c4b8dd4-9sl4b in out of cluster comm: pod \\\"controller-manager-5f4c4b8dd4-9sl4b\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" podUID="03ca839d-b5ba-4b55-87b3-48abca197163" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.040402 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.090017 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.134429 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.164782 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.197923 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.227044 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.293909 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.506297 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.525749 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.584660 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.621929 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.632322 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.745237 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.859126 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.872872 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.977113 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 14:11:13 crc kubenswrapper[4907]: I0313 14:11:13.993914 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.021235 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.049114 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.233506 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.260650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.294789 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.296669 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.413356 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.415514 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.461650 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.490782 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.598728 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.618280 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.657615 4907 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.657940 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4" gracePeriod=5 Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.700326 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.746459 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.761322 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.846540 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.863606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.870912 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.881372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.910621 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.942074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 14:11:14 crc kubenswrapper[4907]: I0313 14:11:14.972004 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.080213 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.111398 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.156561 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.224266 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.265269 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.283335 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.387635 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.387849 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.406396 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.479217 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.479630 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.480957 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.766632 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.773754 4907 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.787283 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.825081 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.898787 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.926795 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.927494 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.931188 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 14:11:15 crc kubenswrapper[4907]: I0313 14:11:15.971633 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.010562 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.031998 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.130230 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.248607 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.290010 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.324510 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.393523 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.459815 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.593924 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.650282 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.656244 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.882919 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 14:11:16 crc kubenswrapper[4907]: I0313 14:11:16.895967 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.024061 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.039914 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.293662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.313386 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.315191 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.390267 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.472966 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.649465 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.747068 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 14:11:17 crc kubenswrapper[4907]: I0313 14:11:17.923872 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.034281 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.105934 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.167740 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.240889 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.409456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.414809 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.469139 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.499685 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 14:11:18 crc kubenswrapper[4907]: I0313 14:11:18.741176 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.059500 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.075649 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.256190 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.394705 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.458135 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.603164 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 14:11:19 crc kubenswrapper[4907]: I0313 14:11:19.897333 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.050799 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.154625 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.154695 4907 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4" exitCode=137 Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.265244 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.265353 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310571 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310629 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310662 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310686 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310757 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310823 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310896 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.310963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311177 4907 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311204 4907 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311226 4907 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.311244 4907 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.413084 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.428998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:11:20 crc kubenswrapper[4907]: I0313 14:11:20.515337 4907 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.101700 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.166163 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.166251 4907 scope.go:117] "RemoveContainer" containerID="d4f9add6c133d321c4423223d8d40952daa019ea189e4fd19cc3ba6806cde1d4" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.166409 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.408731 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.672861 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 14:11:21 crc kubenswrapper[4907]: I0313 14:11:21.793240 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 13 14:11:23 crc kubenswrapper[4907]: I0313 14:11:23.781936 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:23 crc kubenswrapper[4907]: I0313 14:11:23.782546 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:24 crc kubenswrapper[4907]: I0313 14:11:24.326606 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp"] Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.196991 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" event={"ID":"8746df2e-524e-42a7-9ca5-c5300c0a1a09","Type":"ContainerStarted","Data":"27be89dc86edaee6c3a0b2c99e08d3ec32bb36d3d79ee3a0893758666c47e531"} Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.197542 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.197565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" event={"ID":"8746df2e-524e-42a7-9ca5-c5300c0a1a09","Type":"ContainerStarted","Data":"e4b37a2df1824ba52cb67284261294565312dc7c7f4b756f55bce06a3ca979d2"} Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.208116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" Mar 13 14:11:25 crc kubenswrapper[4907]: I0313 14:11:25.233091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7c65b78cb7-ld7fp" podStartSLOduration=53.233048975 podStartE2EDuration="53.233048975s" podCreationTimestamp="2026-03-13 14:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:25.224326203 +0000 UTC m=+384.124113912" watchObservedRunningTime="2026-03-13 14:11:25.233048975 +0000 UTC m=+384.132836704" Mar 13 14:11:26 crc kubenswrapper[4907]: I0313 14:11:26.782330 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:26 crc kubenswrapper[4907]: I0313 14:11:26.783038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.029467 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc"] Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.217200 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" event={"ID":"a704bdfc-46ff-4a8a-8d31-21a140b49c99","Type":"ContainerStarted","Data":"44caeeeda87dc10eb9a28d60f555da1b49720cc085c4d59ad8abb1edafade068"} Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.218029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" event={"ID":"a704bdfc-46ff-4a8a-8d31-21a140b49c99","Type":"ContainerStarted","Data":"daed4ef5759632eb1579bb1e0cd602bb20107fbcc89063bcb1d061a61c877b80"} Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.250600 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" podStartSLOduration=55.250561214 podStartE2EDuration="55.250561214s" podCreationTimestamp="2026-03-13 14:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:27.248256184 +0000 UTC m=+386.148043903" watchObservedRunningTime="2026-03-13 14:11:27.250561214 +0000 UTC m=+386.150348933" Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.782043 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:27 crc kubenswrapper[4907]: I0313 14:11:27.782726 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:28 crc kubenswrapper[4907]: I0313 14:11:28.225162 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:28 crc kubenswrapper[4907]: I0313 14:11:28.233055 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c879c7f8d-2mxxc" Mar 13 14:11:28 crc kubenswrapper[4907]: I0313 14:11:28.331248 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b"] Mar 13 14:11:28 crc kubenswrapper[4907]: W0313 14:11:28.339508 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03ca839d_b5ba_4b55_87b3_48abca197163.slice/crio-f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5 WatchSource:0}: Error finding container f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5: Status 404 returned error can't find the container with id f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5 Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.234288 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" event={"ID":"03ca839d-b5ba-4b55-87b3-48abca197163","Type":"ContainerStarted","Data":"1b8fbae953b8242b87d46728d1b3fbed97457d9be26f86a62170e7caaa324132"} Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.234808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" event={"ID":"03ca839d-b5ba-4b55-87b3-48abca197163","Type":"ContainerStarted","Data":"f2f5e884109c6f169297c52fed01016438e86d1f9eba43973bb35837fea367d5"} Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.258312 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" podStartSLOduration=58.258284449 podStartE2EDuration="58.258284449s" podCreationTimestamp="2026-03-13 14:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:11:29.257460393 +0000 UTC m=+388.157248142" watchObservedRunningTime="2026-03-13 14:11:29.258284449 +0000 UTC m=+388.158072158" Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.687520 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:29 crc kubenswrapper[4907]: I0313 14:11:29.695760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f4c4b8dd4-9sl4b" Mar 13 14:11:40 crc kubenswrapper[4907]: I0313 14:11:40.331411 4907 generic.go:334] "Generic (PLEG): container finished" podID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerID="48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b" exitCode=0 Mar 13 14:11:40 crc kubenswrapper[4907]: I0313 14:11:40.331488 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerDied","Data":"48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b"} Mar 13 14:11:40 crc kubenswrapper[4907]: I0313 14:11:40.333506 4907 scope.go:117] "RemoveContainer" containerID="48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b" Mar 13 14:11:41 crc kubenswrapper[4907]: I0313 14:11:41.345495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerStarted","Data":"a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a"} Mar 13 14:11:41 crc kubenswrapper[4907]: I0313 14:11:41.346358 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:11:41 crc kubenswrapper[4907]: I0313 14:11:41.351963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.186222 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:12:00 crc kubenswrapper[4907]: E0313 14:12:00.187222 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.187238 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.187374 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.187823 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.191143 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.191191 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.193534 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.207965 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.290345 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"auto-csr-approver-29556852-bg4rn\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.391090 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"auto-csr-approver-29556852-bg4rn\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.413764 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"auto-csr-approver-29556852-bg4rn\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:00 crc kubenswrapper[4907]: I0313 14:12:00.508625 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:01 crc kubenswrapper[4907]: I0313 14:12:01.012543 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:12:01 crc kubenswrapper[4907]: I0313 14:12:01.469120 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" event={"ID":"5cd3c3cd-9c71-4d90-a5cb-800de65b86de","Type":"ContainerStarted","Data":"4d8bca074dafc377113ee92ef77c5daac837754efb134951fc4f369ffa3d81f3"} Mar 13 14:12:03 crc kubenswrapper[4907]: I0313 14:12:03.486142 4907 generic.go:334] "Generic (PLEG): container finished" podID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerID="cbcf7b779b5c51c01226dc524be3c2eb43327ef749d0c0644c47a0b6f9140e3e" exitCode=0 Mar 13 14:12:03 crc kubenswrapper[4907]: I0313 14:12:03.486227 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" event={"ID":"5cd3c3cd-9c71-4d90-a5cb-800de65b86de","Type":"ContainerDied","Data":"cbcf7b779b5c51c01226dc524be3c2eb43327ef749d0c0644c47a0b6f9140e3e"} Mar 13 14:12:04 crc kubenswrapper[4907]: I0313 14:12:04.888539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:04 crc kubenswrapper[4907]: I0313 14:12:04.970302 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") pod \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\" (UID: \"5cd3c3cd-9c71-4d90-a5cb-800de65b86de\") " Mar 13 14:12:04 crc kubenswrapper[4907]: I0313 14:12:04.977847 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc" (OuterVolumeSpecName: "kube-api-access-fxcqc") pod "5cd3c3cd-9c71-4d90-a5cb-800de65b86de" (UID: "5cd3c3cd-9c71-4d90-a5cb-800de65b86de"). InnerVolumeSpecName "kube-api-access-fxcqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.071742 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxcqc\" (UniqueName: \"kubernetes.io/projected/5cd3c3cd-9c71-4d90-a5cb-800de65b86de-kube-api-access-fxcqc\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.501909 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" event={"ID":"5cd3c3cd-9c71-4d90-a5cb-800de65b86de","Type":"ContainerDied","Data":"4d8bca074dafc377113ee92ef77c5daac837754efb134951fc4f369ffa3d81f3"} Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.502363 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d8bca074dafc377113ee92ef77c5daac837754efb134951fc4f369ffa3d81f3" Mar 13 14:12:05 crc kubenswrapper[4907]: I0313 14:12:05.501987 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556852-bg4rn" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.881213 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m64n6"] Mar 13 14:12:17 crc kubenswrapper[4907]: E0313 14:12:17.882490 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerName="oc" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.882518 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerName="oc" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.882765 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" containerName="oc" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.883552 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:17 crc kubenswrapper[4907]: I0313 14:12:17.894192 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m64n6"] Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.004992 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005067 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-registry-certificates\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005139 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e905031a-4656-47cf-8f91-92feb42bd418-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7mcj\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-kube-api-access-z7mcj\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e905031a-4656-47cf-8f91-92feb42bd418-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-bound-sa-token\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005274 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-registry-tls\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.005454 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-trusted-ca\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.041699 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.041814 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.047133 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.106511 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-registry-certificates\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107223 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e905031a-4656-47cf-8f91-92feb42bd418-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7mcj\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-kube-api-access-z7mcj\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107666 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e905031a-4656-47cf-8f91-92feb42bd418-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.107910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-bound-sa-token\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.108040 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-registry-tls\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.108333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-trusted-ca\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.110011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-registry-certificates\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.110937 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e905031a-4656-47cf-8f91-92feb42bd418-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.113911 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e905031a-4656-47cf-8f91-92feb42bd418-trusted-ca\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.115767 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e905031a-4656-47cf-8f91-92feb42bd418-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.115979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-registry-tls\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.127571 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-bound-sa-token\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.130837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7mcj\" (UniqueName: \"kubernetes.io/projected/e905031a-4656-47cf-8f91-92feb42bd418-kube-api-access-z7mcj\") pod \"image-registry-66df7c8f76-m64n6\" (UID: \"e905031a-4656-47cf-8f91-92feb42bd418\") " pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.210567 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:18 crc kubenswrapper[4907]: I0313 14:12:18.751363 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m64n6"] Mar 13 14:12:18 crc kubenswrapper[4907]: W0313 14:12:18.765324 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode905031a_4656_47cf_8f91_92feb42bd418.slice/crio-9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c WatchSource:0}: Error finding container 9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c: Status 404 returned error can't find the container with id 9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.606420 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" event={"ID":"e905031a-4656-47cf-8f91-92feb42bd418","Type":"ContainerStarted","Data":"d2b2e692b861662142157ab27f24d6a4e1fd8452f651b0b9aff7165302235cf1"} Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.606911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" event={"ID":"e905031a-4656-47cf-8f91-92feb42bd418","Type":"ContainerStarted","Data":"9e3ef090e9727e8123ae7bc6389b4e34aa9b15f035fbad5f010800f5620b937c"} Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.607067 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:19 crc kubenswrapper[4907]: I0313 14:12:19.637731 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" podStartSLOduration=2.637699189 podStartE2EDuration="2.637699189s" podCreationTimestamp="2026-03-13 14:12:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:12:19.632326635 +0000 UTC m=+438.532114374" watchObservedRunningTime="2026-03-13 14:12:19.637699189 +0000 UTC m=+438.537486888" Mar 13 14:12:38 crc kubenswrapper[4907]: I0313 14:12:38.219324 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-m64n6" Mar 13 14:12:38 crc kubenswrapper[4907]: I0313 14:12:38.299414 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.964003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.964437 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.965381 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:45 crc kubenswrapper[4907]: I0313 14:12:45.972530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.083692 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 14:12:46 crc kubenswrapper[4907]: W0313 14:12:46.552585 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1 WatchSource:0}: Error finding container 5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1: Status 404 returned error can't find the container with id 5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1 Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.787905 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"eef0b64c634c7fd2961d0bde201f0af44757923cc03467a3eafaed052eaee559"} Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.788184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5d3e13e6ce85bd005c8351c1e2a74195a6cc191dcea89de87006f40de5e7faf1"} Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.976461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.976536 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.983709 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:46 crc kubenswrapper[4907]: I0313 14:12:46.983724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.083012 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.182966 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 14:12:47 crc kubenswrapper[4907]: W0313 14:12:47.622922 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc WatchSource:0}: Error finding container b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc: Status 404 returned error can't find the container with id b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.794805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2c35eeee99bdfd18cc00cf2324a49b157be30fef45a8552698e3dd13c84db49b"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.794870 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b589f6199d516fdd4b4f70c3b9a2d237feaf198d2736eb5b0f1edaee3f88ecbc"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.800975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7f17450bbf9f2a07a348d0fc0308410472e1ad570e624a35de5fcbe6939b6ba2"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.801037 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9ba1178efe0855315d30bc9b7bd8d2497870c71f7a0bf4c277765b33fd0a3dd8"} Mar 13 14:12:47 crc kubenswrapper[4907]: I0313 14:12:47.801476 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:12:48 crc kubenswrapper[4907]: I0313 14:12:48.041291 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:12:48 crc kubenswrapper[4907]: I0313 14:12:48.041396 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.526725 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.527750 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-42kpk" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" containerID="cri-o://779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.549658 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.550257 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v6msv" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" containerID="cri-o://6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.556185 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.556581 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" containerID="cri-o://a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.566057 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.566326 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jk5nw" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" containerID="cri-o://de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.583613 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h7vlt"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.586422 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.595353 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.595918 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fmzmc" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" containerID="cri-o://c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049" gracePeriod=30 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.600610 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h7vlt"] Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.685082 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.685133 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.685163 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-266nc\" (UniqueName: \"kubernetes.io/projected/05325ebe-cad6-417a-b1fb-af1efdf24b76-kube-api-access-266nc\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.786019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.786115 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.786145 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-266nc\" (UniqueName: \"kubernetes.io/projected/05325ebe-cad6-417a-b1fb-af1efdf24b76-kube-api-access-266nc\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.788340 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.799938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/05325ebe-cad6-417a-b1fb-af1efdf24b76-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.820581 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-266nc\" (UniqueName: \"kubernetes.io/projected/05325ebe-cad6-417a-b1fb-af1efdf24b76-kube-api-access-266nc\") pod \"marketplace-operator-79b997595-h7vlt\" (UID: \"05325ebe-cad6-417a-b1fb-af1efdf24b76\") " pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.859117 4907 generic.go:334] "Generic (PLEG): container finished" podID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerID="779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.859186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.863895 4907 generic.go:334] "Generic (PLEG): container finished" podID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerID="a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.864046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerDied","Data":"a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.864119 4907 scope.go:117] "RemoveContainer" containerID="48afa92be95670a8e5bf5c9bca4dd6ed4ea5b0ffd5c088a0a66004cb2137ee6b" Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.867366 4907 generic.go:334] "Generic (PLEG): container finished" podID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerID="de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.867415 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.870986 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerID="c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.871133 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.873207 4907 generic.go:334] "Generic (PLEG): container finished" podID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerID="6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f" exitCode=0 Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.873237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f"} Mar 13 14:12:53 crc kubenswrapper[4907]: I0313 14:12:53.924777 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.036679 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.090620 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") pod \"84c2ada5-ceab-4327-802e-9ae459ac814d\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.090672 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") pod \"84c2ada5-ceab-4327-802e-9ae459ac814d\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.090750 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") pod \"84c2ada5-ceab-4327-802e-9ae459ac814d\" (UID: \"84c2ada5-ceab-4327-802e-9ae459ac814d\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.092006 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities" (OuterVolumeSpecName: "utilities") pod "84c2ada5-ceab-4327-802e-9ae459ac814d" (UID: "84c2ada5-ceab-4327-802e-9ae459ac814d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.098443 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c" (OuterVolumeSpecName: "kube-api-access-k6s9c") pod "84c2ada5-ceab-4327-802e-9ae459ac814d" (UID: "84c2ada5-ceab-4327-802e-9ae459ac814d"). InnerVolumeSpecName "kube-api-access-k6s9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.108834 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.116833 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.117558 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.158645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84c2ada5-ceab-4327-802e-9ae459ac814d" (UID: "84c2ada5-ceab-4327-802e-9ae459ac814d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.164640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192151 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") pod \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192208 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") pod \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192327 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") pod \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192374 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") pod \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") pod \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192424 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") pod \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192442 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") pod \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") pod \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") pod \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\" (UID: \"8adb3c1c-bacd-4cca-9796-7ca96624e9f6\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192498 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") pod \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\" (UID: \"d342a6a2-43dd-4d15-8aff-f60f9a3383bf\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") pod \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\" (UID: \"0290eb6f-0be7-4340-8ba8-ed6ea0662c33\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192538 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") pod \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\" (UID: \"ec3466c2-4954-459d-88d8-e8c6cbf309e8\") " Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192802 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192815 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84c2ada5-ceab-4327-802e-9ae459ac814d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.192826 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6s9c\" (UniqueName: \"kubernetes.io/projected/84c2ada5-ceab-4327-802e-9ae459ac814d-kube-api-access-k6s9c\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.196044 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities" (OuterVolumeSpecName: "utilities") pod "d342a6a2-43dd-4d15-8aff-f60f9a3383bf" (UID: "d342a6a2-43dd-4d15-8aff-f60f9a3383bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.196436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0290eb6f-0be7-4340-8ba8-ed6ea0662c33" (UID: "0290eb6f-0be7-4340-8ba8-ed6ea0662c33"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.197137 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities" (OuterVolumeSpecName: "utilities") pod "ec3466c2-4954-459d-88d8-e8c6cbf309e8" (UID: "ec3466c2-4954-459d-88d8-e8c6cbf309e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.197363 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities" (OuterVolumeSpecName: "utilities") pod "8adb3c1c-bacd-4cca-9796-7ca96624e9f6" (UID: "8adb3c1c-bacd-4cca-9796-7ca96624e9f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.200800 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb" (OuterVolumeSpecName: "kube-api-access-48ddb") pod "ec3466c2-4954-459d-88d8-e8c6cbf309e8" (UID: "ec3466c2-4954-459d-88d8-e8c6cbf309e8"). InnerVolumeSpecName "kube-api-access-48ddb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.201333 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0290eb6f-0be7-4340-8ba8-ed6ea0662c33" (UID: "0290eb6f-0be7-4340-8ba8-ed6ea0662c33"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.202024 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6" (OuterVolumeSpecName: "kube-api-access-qppp6") pod "0290eb6f-0be7-4340-8ba8-ed6ea0662c33" (UID: "0290eb6f-0be7-4340-8ba8-ed6ea0662c33"). InnerVolumeSpecName "kube-api-access-qppp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.202570 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t" (OuterVolumeSpecName: "kube-api-access-msd8t") pod "d342a6a2-43dd-4d15-8aff-f60f9a3383bf" (UID: "d342a6a2-43dd-4d15-8aff-f60f9a3383bf"). InnerVolumeSpecName "kube-api-access-msd8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.203782 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs" (OuterVolumeSpecName: "kube-api-access-2x9qs") pod "8adb3c1c-bacd-4cca-9796-7ca96624e9f6" (UID: "8adb3c1c-bacd-4cca-9796-7ca96624e9f6"). InnerVolumeSpecName "kube-api-access-2x9qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.227498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d342a6a2-43dd-4d15-8aff-f60f9a3383bf" (UID: "d342a6a2-43dd-4d15-8aff-f60f9a3383bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.262651 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8adb3c1c-bacd-4cca-9796-7ca96624e9f6" (UID: "8adb3c1c-bacd-4cca-9796-7ca96624e9f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293542 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293566 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48ddb\" (UniqueName: \"kubernetes.io/projected/ec3466c2-4954-459d-88d8-e8c6cbf309e8-kube-api-access-48ddb\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293579 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293587 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293595 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293603 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293611 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x9qs\" (UniqueName: \"kubernetes.io/projected/8adb3c1c-bacd-4cca-9796-7ca96624e9f6-kube-api-access-2x9qs\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293619 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msd8t\" (UniqueName: \"kubernetes.io/projected/d342a6a2-43dd-4d15-8aff-f60f9a3383bf-kube-api-access-msd8t\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293628 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293636 4907 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.293648 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qppp6\" (UniqueName: \"kubernetes.io/projected/0290eb6f-0be7-4340-8ba8-ed6ea0662c33-kube-api-access-qppp6\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.317020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec3466c2-4954-459d-88d8-e8c6cbf309e8" (UID: "ec3466c2-4954-459d-88d8-e8c6cbf309e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.395672 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec3466c2-4954-459d-88d8-e8c6cbf309e8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.553254 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h7vlt"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.881777 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-42kpk" event={"ID":"8adb3c1c-bacd-4cca-9796-7ca96624e9f6","Type":"ContainerDied","Data":"04064aa9c533eaef085c99222c5a674a4c851c29b8b6bd8d0141f59c380540da"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.881826 4907 scope.go:117] "RemoveContainer" containerID="779541283cf16b48730f8ef9601537920e362fa4a7dbb572a3600beba4418986" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.881952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-42kpk" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.890495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" event={"ID":"0290eb6f-0be7-4340-8ba8-ed6ea0662c33","Type":"ContainerDied","Data":"97e42fd7291025df6972a91466dcaf06ac63f386d1c5751baba2fa6aca2cf00a"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.890592 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5stnc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.897758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jk5nw" event={"ID":"d342a6a2-43dd-4d15-8aff-f60f9a3383bf","Type":"ContainerDied","Data":"4fd25f47bfaf1894861177f77553de05bdcc490dd081b002f769378e51f3d5fb"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.897799 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jk5nw" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.901532 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmzmc" event={"ID":"ec3466c2-4954-459d-88d8-e8c6cbf309e8","Type":"ContainerDied","Data":"33655dcb12d3167aaf6db2e72d760da63675955f0f25e1e18f759d3dc24a92c2"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.901599 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmzmc" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.909957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6msv" event={"ID":"84c2ada5-ceab-4327-802e-9ae459ac814d","Type":"ContainerDied","Data":"2b528cfb1a9dfb9a8a4cae32fcb0401d81595581c8b3e1b33eacedab76ae1cd0"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.909982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6msv" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.911372 4907 scope.go:117] "RemoveContainer" containerID="059ebb72320a2144f692825d22d2d54a1e276a63cc41c87a83327001d7f8b539" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.914698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" event={"ID":"05325ebe-cad6-417a-b1fb-af1efdf24b76","Type":"ContainerStarted","Data":"0a42b4b4dddcb9d47eabfa57af67fd59f71354ad57a5dd141a5e496ca7bcc386"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.915023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" event={"ID":"05325ebe-cad6-417a-b1fb-af1efdf24b76","Type":"ContainerStarted","Data":"6c4482ad907f1cc1aba645f7a3ee94d3b6cf5739b1adec2a4a8c45af04f764e2"} Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.915903 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.917375 4907 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-h7vlt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.917411 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" podUID="05325ebe-cad6-417a-b1fb-af1efdf24b76" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.926965 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.932732 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-42kpk"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.951816 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.955338 4907 scope.go:117] "RemoveContainer" containerID="0d444f1a14d20e1e6a2efc71bfbe0eb77305492df11ed2559d59993943816ad1" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.961159 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5stnc"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.972036 4907 scope.go:117] "RemoveContainer" containerID="a4c4a9be5faa2f17ce6e81edeed8ac29998640aca39360db9beeef84d8b0752a" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.975953 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" podStartSLOduration=1.975937745 podStartE2EDuration="1.975937745s" podCreationTimestamp="2026-03-13 14:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:12:54.975036741 +0000 UTC m=+473.874824430" watchObservedRunningTime="2026-03-13 14:12:54.975937745 +0000 UTC m=+473.875725444" Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.997167 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:12:54 crc kubenswrapper[4907]: I0313 14:12:54.998757 4907 scope.go:117] "RemoveContainer" containerID="de346bdbb6259a737c213b9565b9465613c835283a744df647429b69400db2ec" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.002068 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jk5nw"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.012740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.014052 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v6msv"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.020523 4907 scope.go:117] "RemoveContainer" containerID="2ba2c849db65019cfc185d5e8fb816c1ea6e14d84b987e50fe979ce8ace8b5fb" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.026479 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.029638 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fmzmc"] Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.034234 4907 scope.go:117] "RemoveContainer" containerID="167f39bf875dea575d6960ea51e3427665c28d2ee610e574bf41709836edc603" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.055229 4907 scope.go:117] "RemoveContainer" containerID="c60d7cfe210d5d163b6c716dca40da31b6a520923cab2dc2f6be072d3c2f2049" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.087252 4907 scope.go:117] "RemoveContainer" containerID="a06107e49962bc7b9a8f0ee9719c73d09f58fc4ae0da4f8a46e5df6c62712945" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.112237 4907 scope.go:117] "RemoveContainer" containerID="5129da37c4baac30c1ed93e43558cb65af057e72717f5cffd7b1409299226e40" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.135235 4907 scope.go:117] "RemoveContainer" containerID="6e642c222cc08b2da3f2c537efb7c366f76e02de754a7e9abbd7e89b0a44893f" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.160048 4907 scope.go:117] "RemoveContainer" containerID="701b8cbaf2465addef85f7946e5e3a30ad0261ab612755a656d7086a119dc2d2" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.177895 4907 scope.go:117] "RemoveContainer" containerID="0bd7973d57fde27ab0e7adf532d86c90333b4abd96119d9c054a7fb89cbb6dd6" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.790911 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" path="/var/lib/kubelet/pods/0290eb6f-0be7-4340-8ba8-ed6ea0662c33/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.791571 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" path="/var/lib/kubelet/pods/84c2ada5-ceab-4327-802e-9ae459ac814d/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.792405 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" path="/var/lib/kubelet/pods/8adb3c1c-bacd-4cca-9796-7ca96624e9f6/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.793753 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" path="/var/lib/kubelet/pods/d342a6a2-43dd-4d15-8aff-f60f9a3383bf/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.794531 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" path="/var/lib/kubelet/pods/ec3466c2-4954-459d-88d8-e8c6cbf309e8/volumes" Mar 13 14:12:55 crc kubenswrapper[4907]: I0313 14:12:55.934652 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-h7vlt" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.548978 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzhn"] Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549176 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549189 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549200 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549206 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549215 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549220 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549229 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549235 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549243 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549248 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549256 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549262 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549271 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549277 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549288 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549294 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549316 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549322 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-content" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549329 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549335 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549342 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549348 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549358 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549363 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549372 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: E0313 14:12:56.549385 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549391 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="extract-utilities" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549469 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="84c2ada5-ceab-4327-802e-9ae459ac814d" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549479 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d342a6a2-43dd-4d15-8aff-f60f9a3383bf" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549493 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec3466c2-4954-459d-88d8-e8c6cbf309e8" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549501 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8adb3c1c-bacd-4cca-9796-7ca96624e9f6" containerName="registry-server" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549510 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.549517 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0290eb6f-0be7-4340-8ba8-ed6ea0662c33" containerName="marketplace-operator" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.550234 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.554124 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.570054 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzhn"] Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.726275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l4kd\" (UniqueName: \"kubernetes.io/projected/a4b992e7-af9d-44a5-91ea-4942ad99bce2-kube-api-access-6l4kd\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.726437 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-utilities\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.726477 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-catalog-content\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-utilities\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829170 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-catalog-content\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829250 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l4kd\" (UniqueName: \"kubernetes.io/projected/a4b992e7-af9d-44a5-91ea-4942ad99bce2-kube-api-access-6l4kd\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829634 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-utilities\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.829874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4b992e7-af9d-44a5-91ea-4942ad99bce2-catalog-content\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.851420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l4kd\" (UniqueName: \"kubernetes.io/projected/a4b992e7-af9d-44a5-91ea-4942ad99bce2-kube-api-access-6l4kd\") pod \"redhat-marketplace-8dzhn\" (UID: \"a4b992e7-af9d-44a5-91ea-4942ad99bce2\") " pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:56 crc kubenswrapper[4907]: I0313 14:12:56.884958 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.145045 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z7sd2"] Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.146834 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.164228 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.166839 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7sd2"] Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.335968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-utilities\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.336066 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z568x\" (UniqueName: \"kubernetes.io/projected/9d354400-9ba4-45b0-a866-09da4921210a-kube-api-access-z568x\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.336112 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-catalog-content\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.355406 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dzhn"] Mar 13 14:12:57 crc kubenswrapper[4907]: W0313 14:12:57.360648 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4b992e7_af9d_44a5_91ea_4942ad99bce2.slice/crio-d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538 WatchSource:0}: Error finding container d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538: Status 404 returned error can't find the container with id d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538 Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.437597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-utilities\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.437666 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z568x\" (UniqueName: \"kubernetes.io/projected/9d354400-9ba4-45b0-a866-09da4921210a-kube-api-access-z568x\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.437697 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-catalog-content\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.438302 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-catalog-content\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.438302 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d354400-9ba4-45b0-a866-09da4921210a-utilities\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.467629 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z568x\" (UniqueName: \"kubernetes.io/projected/9d354400-9ba4-45b0-a866-09da4921210a-kube-api-access-z568x\") pod \"redhat-operators-z7sd2\" (UID: \"9d354400-9ba4-45b0-a866-09da4921210a\") " pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.490985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.959458 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7sd2"] Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.968235 4907 generic.go:334] "Generic (PLEG): container finished" podID="a4b992e7-af9d-44a5-91ea-4942ad99bce2" containerID="8ff120885864b840470bc227f13746fd7a689b3d184728b2e61e8f3a9ba2cdee" exitCode=0 Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.968327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerDied","Data":"8ff120885864b840470bc227f13746fd7a689b3d184728b2e61e8f3a9ba2cdee"} Mar 13 14:12:57 crc kubenswrapper[4907]: I0313 14:12:57.968383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerStarted","Data":"d3f829163785e2f0fbd5b9a1c153ca084ff17e459a2df81a1f77d5c94887b538"} Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.953916 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.958251 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.962381 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.963643 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.985555 4907 generic.go:334] "Generic (PLEG): container finished" podID="9d354400-9ba4-45b0-a866-09da4921210a" containerID="d6c37d7b57c4e69c9a3f6b656d4baa52460de34948395e9868d67ec94be94892" exitCode=0 Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.985623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerDied","Data":"d6c37d7b57c4e69c9a3f6b656d4baa52460de34948395e9868d67ec94be94892"} Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.985648 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerStarted","Data":"47ed7182a9338cb13165cfe12b9a8fb553637da7ef787dcade4cffd32ec34556"} Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.989988 4907 generic.go:334] "Generic (PLEG): container finished" podID="a4b992e7-af9d-44a5-91ea-4942ad99bce2" containerID="547879c2a6adc0d798d067b0e8539fbf29b569018429785d08c1af1bcb78f58c" exitCode=0 Mar 13 14:12:58 crc kubenswrapper[4907]: I0313 14:12:58.990045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerDied","Data":"547879c2a6adc0d798d067b0e8539fbf29b569018429785d08c1af1bcb78f58c"} Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.059184 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.059251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.059277 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.160967 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161128 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161758 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.161843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.183328 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"community-operators-bkqkq\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.284115 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.489107 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:12:59 crc kubenswrapper[4907]: W0313 14:12:59.499098 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f55641f_96d1_4abe_b29e_35c03423c86f.slice/crio-51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08 WatchSource:0}: Error finding container 51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08: Status 404 returned error can't find the container with id 51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08 Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.548596 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-slm6d"] Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.552651 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.554817 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.556660 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slm6d"] Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.670659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-utilities\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.670705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w459\" (UniqueName: \"kubernetes.io/projected/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-kube-api-access-4w459\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.670731 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-catalog-content\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.772449 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-utilities\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.772671 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w459\" (UniqueName: \"kubernetes.io/projected/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-kube-api-access-4w459\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.772753 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-catalog-content\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.773035 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-utilities\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.773305 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-catalog-content\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.790616 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w459\" (UniqueName: \"kubernetes.io/projected/aa7e1b0d-049f-4775-bd6b-d7430126fbd3-kube-api-access-4w459\") pod \"certified-operators-slm6d\" (UID: \"aa7e1b0d-049f-4775-bd6b-d7430126fbd3\") " pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.977993 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.997011 4907 generic.go:334] "Generic (PLEG): container finished" podID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" exitCode=0 Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.997087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946"} Mar 13 14:12:59 crc kubenswrapper[4907]: I0313 14:12:59.997138 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerStarted","Data":"51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08"} Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.000065 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dzhn" event={"ID":"a4b992e7-af9d-44a5-91ea-4942ad99bce2","Type":"ContainerStarted","Data":"61dc0c95d112b04f5718295d9fa9342b536564f7e473174dfc22f62bfc51b010"} Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.004397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerStarted","Data":"dab525c88cc920e8d4d1cece348d2b1fdef07a1d9676ecd45d6453cc3380f24a"} Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.047153 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8dzhn" podStartSLOduration=2.533633002 podStartE2EDuration="4.047134278s" podCreationTimestamp="2026-03-13 14:12:56 +0000 UTC" firstStartedPulling="2026-03-13 14:12:57.980303097 +0000 UTC m=+476.880090786" lastFinishedPulling="2026-03-13 14:12:59.493804363 +0000 UTC m=+478.393592062" observedRunningTime="2026-03-13 14:13:00.046425709 +0000 UTC m=+478.946213398" watchObservedRunningTime="2026-03-13 14:13:00.047134278 +0000 UTC m=+478.946921967" Mar 13 14:13:00 crc kubenswrapper[4907]: I0313 14:13:00.412655 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-slm6d"] Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.009115 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa7e1b0d-049f-4775-bd6b-d7430126fbd3" containerID="94cfc15a8b67ee390217c4c4f7a6e9d139f4f9130b30d47ce32fe764058f9702" exitCode=0 Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.009209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerDied","Data":"94cfc15a8b67ee390217c4c4f7a6e9d139f4f9130b30d47ce32fe764058f9702"} Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.009507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerStarted","Data":"bb680d736d304ce74d4c50a4ab2ec6d4e3e931217becaf9f28120e12c315dcc1"} Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.011320 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerStarted","Data":"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654"} Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.013214 4907 generic.go:334] "Generic (PLEG): container finished" podID="9d354400-9ba4-45b0-a866-09da4921210a" containerID="dab525c88cc920e8d4d1cece348d2b1fdef07a1d9676ecd45d6453cc3380f24a" exitCode=0 Mar 13 14:13:01 crc kubenswrapper[4907]: I0313 14:13:01.014115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerDied","Data":"dab525c88cc920e8d4d1cece348d2b1fdef07a1d9676ecd45d6453cc3380f24a"} Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.041919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7sd2" event={"ID":"9d354400-9ba4-45b0-a866-09da4921210a","Type":"ContainerStarted","Data":"5f9ae23bba94014c66ed4e0accae3b53de94f902d021871cc264533c7fb6c724"} Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.049060 4907 generic.go:334] "Generic (PLEG): container finished" podID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" exitCode=0 Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.049132 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654"} Mar 13 14:13:02 crc kubenswrapper[4907]: I0313 14:13:02.069187 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z7sd2" podStartSLOduration=2.58717776 podStartE2EDuration="5.06916975s" podCreationTimestamp="2026-03-13 14:12:57 +0000 UTC" firstStartedPulling="2026-03-13 14:12:58.988480678 +0000 UTC m=+477.888268367" lastFinishedPulling="2026-03-13 14:13:01.470472658 +0000 UTC m=+480.370260357" observedRunningTime="2026-03-13 14:13:02.069068076 +0000 UTC m=+480.968855795" watchObservedRunningTime="2026-03-13 14:13:02.06916975 +0000 UTC m=+480.968957449" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.057726 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa7e1b0d-049f-4775-bd6b-d7430126fbd3" containerID="bb225b43511b66781eff07b7c196782d39311083eeb0149c3f9bb46c2bce2771" exitCode=0 Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.057786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerDied","Data":"bb225b43511b66781eff07b7c196782d39311083eeb0149c3f9bb46c2bce2771"} Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.061065 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerStarted","Data":"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178"} Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.092316 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bkqkq" podStartSLOduration=2.612751367 podStartE2EDuration="5.092297588s" podCreationTimestamp="2026-03-13 14:12:58 +0000 UTC" firstStartedPulling="2026-03-13 14:12:59.998771559 +0000 UTC m=+478.898559248" lastFinishedPulling="2026-03-13 14:13:02.47831775 +0000 UTC m=+481.378105469" observedRunningTime="2026-03-13 14:13:03.090461857 +0000 UTC m=+481.990249546" watchObservedRunningTime="2026-03-13 14:13:03.092297588 +0000 UTC m=+481.992085277" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.371566 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" containerID="cri-o://e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" gracePeriod=30 Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.789507 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927288 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927314 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927355 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927423 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.927490 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") pod \"9c5d91bb-753e-4554-b7a1-6c85fe323875\" (UID: \"9c5d91bb-753e-4554-b7a1-6c85fe323875\") " Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.928463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.928583 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.932992 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5" (OuterVolumeSpecName: "kube-api-access-jb2r5") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "kube-api-access-jb2r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.933337 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.933668 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.937716 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.949968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:13:03 crc kubenswrapper[4907]: I0313 14:13:03.954268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9c5d91bb-753e-4554-b7a1-6c85fe323875" (UID: "9c5d91bb-753e-4554-b7a1-6c85fe323875"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029028 4907 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029068 4907 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c5d91bb-753e-4554-b7a1-6c85fe323875-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029079 4907 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c5d91bb-753e-4554-b7a1-6c85fe323875-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029091 4907 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029099 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb2r5\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-kube-api-access-jb2r5\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029107 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c5d91bb-753e-4554-b7a1-6c85fe323875-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.029119 4907 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c5d91bb-753e-4554-b7a1-6c85fe323875-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.066930 4907 generic.go:334] "Generic (PLEG): container finished" podID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" exitCode=0 Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.066982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerDied","Data":"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa"} Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.067009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" event={"ID":"9c5d91bb-753e-4554-b7a1-6c85fe323875","Type":"ContainerDied","Data":"b92738be37732b07d185fe67067541e81cf30ad45197ff888383d57285721ac7"} Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.066967 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-q9z8k" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.067023 4907 scope.go:117] "RemoveContainer" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.069336 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-slm6d" event={"ID":"aa7e1b0d-049f-4775-bd6b-d7430126fbd3","Type":"ContainerStarted","Data":"39cddf09ec86749cd87a8e8442276ad94033823f9427afcca89f9cc001c858da"} Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.086511 4907 scope.go:117] "RemoveContainer" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" Mar 13 14:13:04 crc kubenswrapper[4907]: E0313 14:13:04.086917 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa\": container with ID starting with e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa not found: ID does not exist" containerID="e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.086952 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa"} err="failed to get container status \"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa\": rpc error: code = NotFound desc = could not find container \"e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa\": container with ID starting with e2452e8d8fe737b72f44a8aba2c006a14d53bdfda4e4a01534d104020e7dd4aa not found: ID does not exist" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.098771 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-slm6d" podStartSLOduration=2.622345007 podStartE2EDuration="5.098752891s" podCreationTimestamp="2026-03-13 14:12:59 +0000 UTC" firstStartedPulling="2026-03-13 14:13:01.013287096 +0000 UTC m=+479.913074825" lastFinishedPulling="2026-03-13 14:13:03.48969502 +0000 UTC m=+482.389482709" observedRunningTime="2026-03-13 14:13:04.098407802 +0000 UTC m=+482.998195501" watchObservedRunningTime="2026-03-13 14:13:04.098752891 +0000 UTC m=+482.998540600" Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.124030 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:13:04 crc kubenswrapper[4907]: I0313 14:13:04.130540 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-q9z8k"] Mar 13 14:13:05 crc kubenswrapper[4907]: I0313 14:13:05.800434 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" path="/var/lib/kubelet/pods/9c5d91bb-753e-4554-b7a1-6c85fe323875/volumes" Mar 13 14:13:06 crc kubenswrapper[4907]: I0313 14:13:06.886115 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:06 crc kubenswrapper[4907]: I0313 14:13:06.886699 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:06 crc kubenswrapper[4907]: I0313 14:13:06.941318 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:07 crc kubenswrapper[4907]: I0313 14:13:07.154693 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8dzhn" Mar 13 14:13:07 crc kubenswrapper[4907]: I0313 14:13:07.491186 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:07 crc kubenswrapper[4907]: I0313 14:13:07.491238 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:08 crc kubenswrapper[4907]: I0313 14:13:08.534214 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z7sd2" podUID="9d354400-9ba4-45b0-a866-09da4921210a" containerName="registry-server" probeResult="failure" output=< Mar 13 14:13:08 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:13:08 crc kubenswrapper[4907]: > Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.285308 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.285394 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.322967 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.980140 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:09 crc kubenswrapper[4907]: I0313 14:13:09.980197 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:10 crc kubenswrapper[4907]: I0313 14:13:10.014259 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:10 crc kubenswrapper[4907]: I0313 14:13:10.144606 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:13:10 crc kubenswrapper[4907]: I0313 14:13:10.156766 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-slm6d" Mar 13 14:13:17 crc kubenswrapper[4907]: I0313 14:13:17.101089 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 14:13:17 crc kubenswrapper[4907]: I0313 14:13:17.532487 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:17 crc kubenswrapper[4907]: I0313 14:13:17.602617 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z7sd2" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.041931 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.041984 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.042033 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.042614 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.042656 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de" gracePeriod=600 Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.180394 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de" exitCode=0 Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.180502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de"} Mar 13 14:13:18 crc kubenswrapper[4907]: I0313 14:13:18.180577 4907 scope.go:117] "RemoveContainer" containerID="770c882d9e9f9b2e9bd4f135e591952fa15b3898c42826aed8a75d953c9a875b" Mar 13 14:13:19 crc kubenswrapper[4907]: I0313 14:13:19.186924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195"} Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.165501 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:14:00 crc kubenswrapper[4907]: E0313 14:14:00.168340 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.168527 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.168986 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c5d91bb-753e-4554-b7a1-6c85fe323875" containerName="registry" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.169941 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.172794 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.173867 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.173962 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.178970 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.241475 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"auto-csr-approver-29556854-l7l9t\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.343657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"auto-csr-approver-29556854-l7l9t\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.384493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"auto-csr-approver-29556854-l7l9t\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.503847 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:00 crc kubenswrapper[4907]: I0313 14:14:00.993671 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:14:01 crc kubenswrapper[4907]: I0313 14:14:01.460228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" event={"ID":"82df106b-e5b8-4798-942a-999c2ac7a4be","Type":"ContainerStarted","Data":"967dbb55436e1bea2443bf62f90aa9ec79cd3b235fe5f9992286f8657a1468eb"} Mar 13 14:14:02 crc kubenswrapper[4907]: I0313 14:14:02.469423 4907 generic.go:334] "Generic (PLEG): container finished" podID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerID="2fab8f0b7470e3acd4d60ef3af4057ac0e1b8ec4f3de833c7e05ac6623d0bf47" exitCode=0 Mar 13 14:14:02 crc kubenswrapper[4907]: I0313 14:14:02.469782 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" event={"ID":"82df106b-e5b8-4798-942a-999c2ac7a4be","Type":"ContainerDied","Data":"2fab8f0b7470e3acd4d60ef3af4057ac0e1b8ec4f3de833c7e05ac6623d0bf47"} Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.746798 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.892963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") pod \"82df106b-e5b8-4798-942a-999c2ac7a4be\" (UID: \"82df106b-e5b8-4798-942a-999c2ac7a4be\") " Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.900942 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n" (OuterVolumeSpecName: "kube-api-access-nkz6n") pod "82df106b-e5b8-4798-942a-999c2ac7a4be" (UID: "82df106b-e5b8-4798-942a-999c2ac7a4be"). InnerVolumeSpecName "kube-api-access-nkz6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:14:03 crc kubenswrapper[4907]: I0313 14:14:03.995586 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkz6n\" (UniqueName: \"kubernetes.io/projected/82df106b-e5b8-4798-942a-999c2ac7a4be-kube-api-access-nkz6n\") on node \"crc\" DevicePath \"\"" Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.489076 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" event={"ID":"82df106b-e5b8-4798-942a-999c2ac7a4be","Type":"ContainerDied","Data":"967dbb55436e1bea2443bf62f90aa9ec79cd3b235fe5f9992286f8657a1468eb"} Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.489158 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="967dbb55436e1bea2443bf62f90aa9ec79cd3b235fe5f9992286f8657a1468eb" Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.489397 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556854-l7l9t" Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.844963 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:14:04 crc kubenswrapper[4907]: I0313 14:14:04.852307 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556848-v8kv4"] Mar 13 14:14:05 crc kubenswrapper[4907]: I0313 14:14:05.797935 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8867781-179d-47e4-9a55-330deb10d273" path="/var/lib/kubelet/pods/a8867781-179d-47e4-9a55-330deb10d273/volumes" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.155639 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 14:15:00 crc kubenswrapper[4907]: E0313 14:15:00.156980 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerName="oc" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.157341 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerName="oc" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.157616 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" containerName="oc" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.158325 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.164039 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.164864 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.182493 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.320778 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.320860 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.320972 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.422263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.422772 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.423091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.424470 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.429705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.443430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"collect-profiles-29556855-t8w6z\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.530484 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.748542 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.872469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerStarted","Data":"1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff"} Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.872507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerStarted","Data":"e783ab47360b0f678f98f9fecaaf31bb829cbc86f6e08ca80e77430e665ca4a4"} Mar 13 14:15:00 crc kubenswrapper[4907]: I0313 14:15:00.889761 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" podStartSLOduration=0.889742696 podStartE2EDuration="889.742696ms" podCreationTimestamp="2026-03-13 14:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:15:00.88665707 +0000 UTC m=+599.786444759" watchObservedRunningTime="2026-03-13 14:15:00.889742696 +0000 UTC m=+599.789530395" Mar 13 14:15:01 crc kubenswrapper[4907]: I0313 14:15:01.880719 4907 generic.go:334] "Generic (PLEG): container finished" podID="80282185-9b05-4101-b906-e069f57e9bae" containerID="1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff" exitCode=0 Mar 13 14:15:01 crc kubenswrapper[4907]: I0313 14:15:01.880764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerDied","Data":"1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff"} Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.190584 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.359425 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") pod \"80282185-9b05-4101-b906-e069f57e9bae\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.359634 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") pod \"80282185-9b05-4101-b906-e069f57e9bae\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.359724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") pod \"80282185-9b05-4101-b906-e069f57e9bae\" (UID: \"80282185-9b05-4101-b906-e069f57e9bae\") " Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.360698 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume" (OuterVolumeSpecName: "config-volume") pod "80282185-9b05-4101-b906-e069f57e9bae" (UID: "80282185-9b05-4101-b906-e069f57e9bae"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.367751 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "80282185-9b05-4101-b906-e069f57e9bae" (UID: "80282185-9b05-4101-b906-e069f57e9bae"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.368790 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv" (OuterVolumeSpecName: "kube-api-access-hjwzv") pod "80282185-9b05-4101-b906-e069f57e9bae" (UID: "80282185-9b05-4101-b906-e069f57e9bae"). InnerVolumeSpecName "kube-api-access-hjwzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.461412 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80282185-9b05-4101-b906-e069f57e9bae-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.461464 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjwzv\" (UniqueName: \"kubernetes.io/projected/80282185-9b05-4101-b906-e069f57e9bae-kube-api-access-hjwzv\") on node \"crc\" DevicePath \"\"" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.461488 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/80282185-9b05-4101-b906-e069f57e9bae-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.896387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" event={"ID":"80282185-9b05-4101-b906-e069f57e9bae","Type":"ContainerDied","Data":"e783ab47360b0f678f98f9fecaaf31bb829cbc86f6e08ca80e77430e665ca4a4"} Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.896439 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e783ab47360b0f678f98f9fecaaf31bb829cbc86f6e08ca80e77430e665ca4a4" Mar 13 14:15:03 crc kubenswrapper[4907]: I0313 14:15:03.896456 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z" Mar 13 14:15:18 crc kubenswrapper[4907]: I0313 14:15:18.042024 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:15:18 crc kubenswrapper[4907]: I0313 14:15:18.042645 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:15:48 crc kubenswrapper[4907]: I0313 14:15:48.041629 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:15:48 crc kubenswrapper[4907]: I0313 14:15:48.042319 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.139582 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:16:00 crc kubenswrapper[4907]: E0313 14:16:00.140192 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80282185-9b05-4101-b906-e069f57e9bae" containerName="collect-profiles" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.140203 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="80282185-9b05-4101-b906-e069f57e9bae" containerName="collect-profiles" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.140303 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="80282185-9b05-4101-b906-e069f57e9bae" containerName="collect-profiles" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.140605 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.143110 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.144145 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.148488 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.162845 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.282705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"auto-csr-approver-29556856-8grxd\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.384694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"auto-csr-approver-29556856-8grxd\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.419509 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"auto-csr-approver-29556856-8grxd\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.459167 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.709862 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:16:00 crc kubenswrapper[4907]: I0313 14:16:00.719141 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:16:01 crc kubenswrapper[4907]: I0313 14:16:01.434686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerStarted","Data":"c5e66cf058a4411962cd24af99e75819cdb33969ea8b7d7705c9750d7b4c56eb"} Mar 13 14:16:02 crc kubenswrapper[4907]: I0313 14:16:02.200814 4907 scope.go:117] "RemoveContainer" containerID="2e545187909a28cc5890020e1139c1d488e394bab920eab3166b317419887a60" Mar 13 14:16:02 crc kubenswrapper[4907]: I0313 14:16:02.444957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerStarted","Data":"ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd"} Mar 13 14:16:02 crc kubenswrapper[4907]: I0313 14:16:02.464091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556856-8grxd" podStartSLOduration=1.092545248 podStartE2EDuration="2.464065981s" podCreationTimestamp="2026-03-13 14:16:00 +0000 UTC" firstStartedPulling="2026-03-13 14:16:00.718784855 +0000 UTC m=+659.618572554" lastFinishedPulling="2026-03-13 14:16:02.090305598 +0000 UTC m=+660.990093287" observedRunningTime="2026-03-13 14:16:02.45975387 +0000 UTC m=+661.359541559" watchObservedRunningTime="2026-03-13 14:16:02.464065981 +0000 UTC m=+661.363853670" Mar 13 14:16:03 crc kubenswrapper[4907]: I0313 14:16:03.458024 4907 generic.go:334] "Generic (PLEG): container finished" podID="1604551e-a53d-483c-a03e-3c2159768001" containerID="ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd" exitCode=0 Mar 13 14:16:03 crc kubenswrapper[4907]: I0313 14:16:03.458143 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerDied","Data":"ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd"} Mar 13 14:16:04 crc kubenswrapper[4907]: I0313 14:16:04.809461 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:04 crc kubenswrapper[4907]: I0313 14:16:04.973700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") pod \"1604551e-a53d-483c-a03e-3c2159768001\" (UID: \"1604551e-a53d-483c-a03e-3c2159768001\") " Mar 13 14:16:04 crc kubenswrapper[4907]: I0313 14:16:04.981376 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k" (OuterVolumeSpecName: "kube-api-access-6sp5k") pod "1604551e-a53d-483c-a03e-3c2159768001" (UID: "1604551e-a53d-483c-a03e-3c2159768001"). InnerVolumeSpecName "kube-api-access-6sp5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.075632 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sp5k\" (UniqueName: \"kubernetes.io/projected/1604551e-a53d-483c-a03e-3c2159768001-kube-api-access-6sp5k\") on node \"crc\" DevicePath \"\"" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.473285 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556856-8grxd" event={"ID":"1604551e-a53d-483c-a03e-3c2159768001","Type":"ContainerDied","Data":"c5e66cf058a4411962cd24af99e75819cdb33969ea8b7d7705c9750d7b4c56eb"} Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.473374 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5e66cf058a4411962cd24af99e75819cdb33969ea8b7d7705c9750d7b4c56eb" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.473708 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556856-8grxd" Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.880154 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:16:05 crc kubenswrapper[4907]: I0313 14:16:05.885290 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556850-cl45g"] Mar 13 14:16:07 crc kubenswrapper[4907]: I0313 14:16:07.793628 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a1d486-2cf1-4078-8f0f-2dbc1ea97acc" path="/var/lib/kubelet/pods/10a1d486-2cf1-4078-8f0f-2dbc1ea97acc/volumes" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.041526 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.042686 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.042778 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.043834 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.043965 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195" gracePeriod=600 Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.567631 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195" exitCode=0 Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.567722 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195"} Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.568407 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e"} Mar 13 14:16:18 crc kubenswrapper[4907]: I0313 14:16:18.568477 4907 scope.go:117] "RemoveContainer" containerID="93da0fb3f9dc1b1da4e8f0430569f23f1e76f6777c8f829bb761fb1ef946f9de" Mar 13 14:17:02 crc kubenswrapper[4907]: I0313 14:17:02.258144 4907 scope.go:117] "RemoveContainer" containerID="7118e101427b23d94455b072a1fe8c88f9a426be2d07a4a232a1f55839394ad6" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.150157 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:18:00 crc kubenswrapper[4907]: E0313 14:18:00.151087 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1604551e-a53d-483c-a03e-3c2159768001" containerName="oc" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.151108 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1604551e-a53d-483c-a03e-3c2159768001" containerName="oc" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.151277 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1604551e-a53d-483c-a03e-3c2159768001" containerName="oc" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.151939 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.154556 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.155044 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.155349 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.170538 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.249087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"auto-csr-approver-29556858-bk6rj\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.351293 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"auto-csr-approver-29556858-bk6rj\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.380393 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"auto-csr-approver-29556858-bk6rj\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.482441 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:00 crc kubenswrapper[4907]: I0313 14:18:00.767073 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:18:01 crc kubenswrapper[4907]: I0313 14:18:01.452773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" event={"ID":"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a","Type":"ContainerStarted","Data":"6cee4ab91f9e43ac6ed054dabe1750fcb13c6fcca497178aea30982262b12a53"} Mar 13 14:18:02 crc kubenswrapper[4907]: I0313 14:18:02.463580 4907 generic.go:334] "Generic (PLEG): container finished" podID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerID="7377ca75e3097f224f1949f7536a8f7a1c19b44224ea656c9d4dd6e3d82e1652" exitCode=0 Mar 13 14:18:02 crc kubenswrapper[4907]: I0313 14:18:02.463667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" event={"ID":"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a","Type":"ContainerDied","Data":"7377ca75e3097f224f1949f7536a8f7a1c19b44224ea656c9d4dd6e3d82e1652"} Mar 13 14:18:03 crc kubenswrapper[4907]: I0313 14:18:03.795878 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:03 crc kubenswrapper[4907]: I0313 14:18:03.903398 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") pod \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\" (UID: \"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a\") " Mar 13 14:18:03 crc kubenswrapper[4907]: I0313 14:18:03.913676 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw" (OuterVolumeSpecName: "kube-api-access-dkhtw") pod "03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" (UID: "03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a"). InnerVolumeSpecName "kube-api-access-dkhtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.005235 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkhtw\" (UniqueName: \"kubernetes.io/projected/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a-kube-api-access-dkhtw\") on node \"crc\" DevicePath \"\"" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.478719 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" event={"ID":"03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a","Type":"ContainerDied","Data":"6cee4ab91f9e43ac6ed054dabe1750fcb13c6fcca497178aea30982262b12a53"} Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.479379 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cee4ab91f9e43ac6ed054dabe1750fcb13c6fcca497178aea30982262b12a53" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.478801 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556858-bk6rj" Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.877386 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:18:04 crc kubenswrapper[4907]: I0313 14:18:04.884510 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556852-bg4rn"] Mar 13 14:18:05 crc kubenswrapper[4907]: I0313 14:18:05.795203 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd3c3cd-9c71-4d90-a5cb-800de65b86de" path="/var/lib/kubelet/pods/5cd3c3cd-9c71-4d90-a5cb-800de65b86de/volumes" Mar 13 14:18:18 crc kubenswrapper[4907]: I0313 14:18:18.323065 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:18:18 crc kubenswrapper[4907]: I0313 14:18:18.324180 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:18:48 crc kubenswrapper[4907]: I0313 14:18:48.041521 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:18:48 crc kubenswrapper[4907]: I0313 14:18:48.042180 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:19:02 crc kubenswrapper[4907]: I0313 14:19:02.343096 4907 scope.go:117] "RemoveContainer" containerID="cbcf7b779b5c51c01226dc524be3c2eb43327ef749d0c0644c47a0b6f9140e3e" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.042105 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.042753 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.042815 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.043631 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.043732 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e" gracePeriod=600 Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727012 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e" exitCode=0 Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e"} Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727671 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce"} Mar 13 14:19:18 crc kubenswrapper[4907]: I0313 14:19:18.727695 4907 scope.go:117] "RemoveContainer" containerID="c6951e52b25c8a7e922f875d35cd18c0cb0419427475b81395b9907dcedc0195" Mar 13 14:19:20 crc kubenswrapper[4907]: I0313 14:19:20.879383 4907 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.269510 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270371 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" containerID="cri-o://fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270731 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" containerID="cri-o://4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270767 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" containerID="cri-o://a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270840 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" containerID="cri-o://0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270919 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.270957 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" containerID="cri-o://bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.271018 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" containerID="cri-o://3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.314645 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" containerID="cri-o://a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" gracePeriod=30 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.611833 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.615351 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-acl-logging/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.616016 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-controller/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.616453 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678091 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8pphf"] Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678466 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678493 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678531 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678542 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678555 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678563 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678576 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678584 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678622 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kubecfg-setup" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678633 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kubecfg-setup" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678646 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678655 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678664 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678694 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678710 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerName="oc" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678721 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerName="oc" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678733 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678740 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678814 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678906 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678931 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678939 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678949 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678958 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: E0313 14:19:42.678970 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.678979 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679111 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679124 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679136 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" containerName="oc" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679148 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="kube-rbac-proxy-node" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679157 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="northd" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679167 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679178 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="nbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679187 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679196 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679206 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="sbdb" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679217 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovnkube-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679227 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-controller" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.679237 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" containerName="ovn-acl-logging" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.682114 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760238 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760273 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760252 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760295 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760325 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760398 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash" (OuterVolumeSpecName: "host-slash") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760416 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760527 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760556 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760612 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760665 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760694 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760731 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760799 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760823 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760852 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760908 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760929 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.760960 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") pod \"1006376d-4033-454e-8a23-2b67882bdd0b\" (UID: \"1006376d-4033-454e-8a23-2b67882bdd0b\") " Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761067 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761103 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761119 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761139 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log" (OuterVolumeSpecName: "node-log") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket" (OuterVolumeSpecName: "log-socket") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761179 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761221 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761090 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761200 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761524 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761536 4907 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761548 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761567 4907 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761646 4907 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761667 4907 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-log-socket\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761688 4907 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761707 4907 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761729 4907 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761752 4907 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761770 4907 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-node-log\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761789 4907 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761809 4907 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761830 4907 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-host-slash\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761849 4907 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761867 4907 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.761921 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.766090 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.766820 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw" (OuterVolumeSpecName: "kube-api-access-tm9lw") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "kube-api-access-tm9lw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.773326 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1006376d-4033-454e-8a23-2b67882bdd0b" (UID: "1006376d-4033-454e-8a23-2b67882bdd0b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.863756 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.863846 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbxsh\" (UniqueName: \"kubernetes.io/projected/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-kube-api-access-jbxsh\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864647 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-netns\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864688 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-netd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-kubelet\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864876 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-env-overrides\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.864993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-log-socket\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865022 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-script-lib\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865096 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-config\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865214 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-var-lib-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865268 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-etc-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865294 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-node-log\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865352 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-ovn\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-systemd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865465 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-bin\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865514 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-systemd-units\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-slash\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865566 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovn-node-metrics-cert\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865642 4907 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865679 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm9lw\" (UniqueName: \"kubernetes.io/projected/1006376d-4033-454e-8a23-2b67882bdd0b-kube-api-access-tm9lw\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865689 4907 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1006376d-4033-454e-8a23-2b67882bdd0b-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865698 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1006376d-4033-454e-8a23-2b67882bdd0b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.865708 4907 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1006376d-4033-454e-8a23-2b67882bdd0b-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.899110 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovnkube-controller/3.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.901734 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-acl-logging/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902264 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qxlx5_1006376d-4033-454e-8a23-2b67882bdd0b/ovn-controller/0.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902754 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902781 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902789 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902798 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902805 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902811 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" exitCode=0 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902818 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" exitCode=143 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902826 4907 generic.go:334] "Generic (PLEG): container finished" podID="1006376d-4033-454e-8a23-2b67882bdd0b" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" exitCode=143 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902856 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902873 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902897 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902909 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902934 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902945 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902954 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902959 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902964 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902970 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902975 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902980 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902986 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902991 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.902998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903004 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903009 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903105 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903118 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903125 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903131 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903136 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903141 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903147 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903153 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903158 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903176 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903196 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903202 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903209 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903214 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903220 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903225 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903230 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903235 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903240 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903245 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qxlx5" event={"ID":"1006376d-4033-454e-8a23-2b67882bdd0b","Type":"ContainerDied","Data":"4a8bfe852be3826bb96e2a956b4a0cd7eaff7036e060420d181a2514210039c0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903261 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903269 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903274 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903280 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903286 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903291 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903296 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903302 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903308 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.903314 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.904726 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/2.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905479 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905506 4907 generic.go:334] "Generic (PLEG): container finished" podID="64ebd338-0158-44dc-90ab-e7c285e87762" containerID="8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173" exitCode=2 Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905526 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerDied","Data":"8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.905546 4907 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b"} Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.906030 4907 scope.go:117] "RemoveContainer" containerID="8788f4b58c32d23e53448793bceb517c39726e72de35c52e959fc15e062ae173" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.935473 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.957239 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.962441 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qxlx5"] Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-node-log\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967258 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-ovn\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-systemd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967338 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967360 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-node-log\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-ovn\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-bin\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-bin\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967439 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-systemd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967792 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967898 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-systemd-units\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967922 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-slash\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovn-node-metrics-cert\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.967982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbxsh\" (UniqueName: \"kubernetes.io/projected/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-kube-api-access-jbxsh\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-netns\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968056 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-netd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-kubelet\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-env-overrides\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968113 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-log-socket\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968128 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-script-lib\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968160 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-config\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968231 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-var-lib-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-etc-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968346 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-etc-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968370 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-systemd-units\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968393 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-slash\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968859 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-run-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-ovn-kubernetes\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-cni-netd\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968932 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-run-netns\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.968965 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-log-socket\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.969148 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-host-kubelet\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.969239 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-var-lib-openvswitch\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.969562 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-script-lib\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.970002 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovnkube-config\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.974353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-env-overrides\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.976157 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-ovn-node-metrics-cert\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.980996 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.993645 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:42 crc kubenswrapper[4907]: I0313 14:19:42.994280 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbxsh\" (UniqueName: \"kubernetes.io/projected/42f6fac8-4aa0-4cd7-a1a1-c80ced66c462-kube-api-access-jbxsh\") pod \"ovnkube-node-8pphf\" (UID: \"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462\") " pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.006304 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.019797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.021586 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.046185 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: W0313 14:19:43.050469 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42f6fac8_4aa0_4cd7_a1a1_c80ced66c462.slice/crio-b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4 WatchSource:0}: Error finding container b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4: Status 404 returned error can't find the container with id b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4 Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.060761 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.073958 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.105059 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.131285 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.131806 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.131845 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.131873 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.132513 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132540 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132561 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.132919 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132948 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.132972 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.133270 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133298 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133315 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.133519 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133547 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133565 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.133897 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133926 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.133945 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.134394 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134421 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134438 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.134723 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134753 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.134770 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.135052 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135084 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135101 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: E0313 14:19:43.135325 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135368 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135383 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135665 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135690 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135931 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.135953 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136160 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136189 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136406 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136427 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136663 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136685 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136863 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.136902 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137122 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137143 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137394 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137477 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137691 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137729 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.137976 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138000 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138183 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138207 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138398 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138419 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138591 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138611 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138859 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.138897 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139151 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139172 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139373 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139394 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139574 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139595 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139778 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.139798 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140019 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140040 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140230 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140251 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140439 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140463 4907 scope.go:117] "RemoveContainer" containerID="9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140653 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee"} err="failed to get container status \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": rpc error: code = NotFound desc = could not find container \"9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee\": container with ID starting with 9155434e2d8f0daa4504758fb72d6f498da107a64a92a0883ac687ed2af9caee not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.140675 4907 scope.go:117] "RemoveContainer" containerID="4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141348 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751"} err="failed to get container status \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": rpc error: code = NotFound desc = could not find container \"4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751\": container with ID starting with 4d60b5b774f8dd4cee16b9784f80d124640199a85331a2940772029be4884751 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141369 4907 scope.go:117] "RemoveContainer" containerID="a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141568 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395"} err="failed to get container status \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": rpc error: code = NotFound desc = could not find container \"a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395\": container with ID starting with a0ee0049aa4a6d49a6bcf7d347183d9cc8b85171f988662803a005e0cdd97395 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141591 4907 scope.go:117] "RemoveContainer" containerID="0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141816 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2"} err="failed to get container status \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": rpc error: code = NotFound desc = could not find container \"0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2\": container with ID starting with 0fff87d8ea572615eb1329164f9e536e6ee45cd00a9e71be0aa8a9b3ed2190e2 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.141838 4907 scope.go:117] "RemoveContainer" containerID="ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142117 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637"} err="failed to get container status \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": rpc error: code = NotFound desc = could not find container \"ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637\": container with ID starting with ccef92e41501fab91189581d7e8875515c02d637426b329d2fb80a356c75b637 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142138 4907 scope.go:117] "RemoveContainer" containerID="bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142350 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600"} err="failed to get container status \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": rpc error: code = NotFound desc = could not find container \"bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600\": container with ID starting with bf6a3dbaa8dda61f461a906538824c8946d36dc07d4fb95bf4d8420ad0105600 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142371 4907 scope.go:117] "RemoveContainer" containerID="3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142561 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d"} err="failed to get container status \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": rpc error: code = NotFound desc = could not find container \"3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d\": container with ID starting with 3c165053fd9933444b372ff40efe33c0e193a7f39e6f531a4f370e971eb41e6d not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142582 4907 scope.go:117] "RemoveContainer" containerID="fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142765 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0"} err="failed to get container status \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": rpc error: code = NotFound desc = could not find container \"fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0\": container with ID starting with fea41c11bebd2d9d622ed4b8bc50e03a106dadea6ac102cd60512ea2a1f8b8b0 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.142787 4907 scope.go:117] "RemoveContainer" containerID="858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.143046 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279"} err="failed to get container status \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": rpc error: code = NotFound desc = could not find container \"858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279\": container with ID starting with 858a7985907035c41390569f49c1f773ccc85ca0dabeffae2ff0520586b10279 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.143069 4907 scope.go:117] "RemoveContainer" containerID="a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.143262 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886"} err="failed to get container status \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": rpc error: code = NotFound desc = could not find container \"a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886\": container with ID starting with a3d2a347b2f85682f2a54c30e63b7e9c88ffe402d6d71fa7ce6656786f93b886 not found: ID does not exist" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.791831 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1006376d-4033-454e-8a23-2b67882bdd0b" path="/var/lib/kubelet/pods/1006376d-4033-454e-8a23-2b67882bdd0b/volumes" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.915126 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/2.log" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.915906 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/1.log" Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.916107 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xtwf7" event={"ID":"64ebd338-0158-44dc-90ab-e7c285e87762","Type":"ContainerStarted","Data":"569f8a799f7f20725d989134c554036cb2d5eb61b9ee2129f2d173a5c1c3e1a9"} Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.921126 4907 generic.go:334] "Generic (PLEG): container finished" podID="42f6fac8-4aa0-4cd7-a1a1-c80ced66c462" containerID="d106a54791454eb6427df73c45bd36a1109922c53a635e0a8d7cf85779280458" exitCode=0 Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.921212 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerDied","Data":"d106a54791454eb6427df73c45bd36a1109922c53a635e0a8d7cf85779280458"} Mar 13 14:19:43 crc kubenswrapper[4907]: I0313 14:19:43.921256 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"b113dbc4695af8ad071eea34316d474e13189383fd3ee685adb103d09acf57d4"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.933911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"578de0ccce050e57e4f8e80bbe09f048bb5cd0dae436f2ba29311e78a016cc11"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.935079 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"25b13197844fea9f278eff609105bf3c042456ddb283a72136938e344204cf4f"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.935186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"0d0df32e73dca5d01127944f2d46839aeee380a1b2ea551c278b048e37bffe05"} Mar 13 14:19:44 crc kubenswrapper[4907]: I0313 14:19:44.935260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"c23967456e62fa626525f0f5f03be5f1b03daa0c6c67d13556df25e9cf70b1ed"} Mar 13 14:19:45 crc kubenswrapper[4907]: I0313 14:19:45.945541 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"28cd9af2d20dd6887662d65303cd8fba04b00f95081ebede4b2cbadeabf91813"} Mar 13 14:19:45 crc kubenswrapper[4907]: I0313 14:19:45.945917 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"9af887e10ec13a531557ac2b8baedf3614b35ac1b0b62a832b549ded6ce7fd7f"} Mar 13 14:19:47 crc kubenswrapper[4907]: I0313 14:19:47.963302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"347baab8da25f1e8a13e9500f5636a4f8b732c7dd039e0d6df7ce84b3cf22f01"} Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.732144 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.733177 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735227 4907 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2d28j" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735339 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735458 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.735913 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.792226 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.792415 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.792654 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.893962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.894041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.894131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.894549 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.895525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:48 crc kubenswrapper[4907]: I0313 14:19:48.924010 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"crc-storage-crc-wds2z\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: I0313 14:19:49.059175 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111605 4907 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111767 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111812 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:49 crc kubenswrapper[4907]: E0313 14:19:49.111920 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(664f7b52eff02406755a65652f95b07d39e80c06474459a884c4a4c3a60ed0fe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-wds2z" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.734059 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.735383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.736303 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780165 4907 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780261 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780295 4907 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:19:50 crc kubenswrapper[4907]: E0313 14:19:50.780369 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-wds2z_crc-storage(b501aea0-66d1-4655-b67f-c6a3bacb7cdb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-wds2z_crc-storage_b501aea0-66d1-4655-b67f-c6a3bacb7cdb_0(fa9f25ad6b8187a7a88d3dd386755b7b0b93bda2b9cade1fbc1357b71fa54a4b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-wds2z" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.987326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" event={"ID":"42f6fac8-4aa0-4cd7-a1a1-c80ced66c462","Type":"ContainerStarted","Data":"a0f5a9d3eb46fe980c3893c4201a77f3a1ab6fe006f338a8077b3da098ea6914"} Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.987872 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.988091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:50 crc kubenswrapper[4907]: I0313 14:19:50.988220 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:51 crc kubenswrapper[4907]: I0313 14:19:51.040228 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" podStartSLOduration=9.040197703 podStartE2EDuration="9.040197703s" podCreationTimestamp="2026-03-13 14:19:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:19:51.027082755 +0000 UTC m=+889.926870484" watchObservedRunningTime="2026-03-13 14:19:51.040197703 +0000 UTC m=+889.939985432" Mar 13 14:19:51 crc kubenswrapper[4907]: I0313 14:19:51.049633 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:19:51 crc kubenswrapper[4907]: I0313 14:19:51.056619 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.148531 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.152439 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.155249 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.157244 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.157403 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.161833 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.249612 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"auto-csr-approver-29556860-ljdj7\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.351174 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"auto-csr-approver-29556860-ljdj7\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.373476 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"auto-csr-approver-29556860-ljdj7\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.484078 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:00 crc kubenswrapper[4907]: I0313 14:20:00.752183 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:20:00 crc kubenswrapper[4907]: W0313 14:20:00.764092 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod257ee294_461c_4344_9dcb_b44f1497f8bf.slice/crio-4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f WatchSource:0}: Error finding container 4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f: Status 404 returned error can't find the container with id 4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f Mar 13 14:20:01 crc kubenswrapper[4907]: I0313 14:20:01.062010 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerStarted","Data":"4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f"} Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.072055 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerStarted","Data":"dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368"} Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.426114 4907 scope.go:117] "RemoveContainer" containerID="ef632bb17c8f529b35d0882c8d4ba3dc213b0308b46a4ade8426b9c08e64bb2b" Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.781698 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:02 crc kubenswrapper[4907]: I0313 14:20:02.782517 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.094423 4907 generic.go:334] "Generic (PLEG): container finished" podID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerID="dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368" exitCode=0 Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.094545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerDied","Data":"dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368"} Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.101126 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xtwf7_64ebd338-0158-44dc-90ab-e7c285e87762/kube-multus/2.log" Mar 13 14:20:03 crc kubenswrapper[4907]: I0313 14:20:03.239182 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 14:20:03 crc kubenswrapper[4907]: W0313 14:20:03.245973 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb501aea0_66d1_4655_b67f_c6a3bacb7cdb.slice/crio-5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a WatchSource:0}: Error finding container 5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a: Status 404 returned error can't find the container with id 5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.111540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wds2z" event={"ID":"b501aea0-66d1-4655-b67f-c6a3bacb7cdb","Type":"ContainerStarted","Data":"5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a"} Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.556988 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.609923 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") pod \"257ee294-461c-4344-9dcb-b44f1497f8bf\" (UID: \"257ee294-461c-4344-9dcb-b44f1497f8bf\") " Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.616521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l" (OuterVolumeSpecName: "kube-api-access-j2h2l") pod "257ee294-461c-4344-9dcb-b44f1497f8bf" (UID: "257ee294-461c-4344-9dcb-b44f1497f8bf"). InnerVolumeSpecName "kube-api-access-j2h2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.711519 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2h2l\" (UniqueName: \"kubernetes.io/projected/257ee294-461c-4344-9dcb-b44f1497f8bf-kube-api-access-j2h2l\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.872805 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:20:04 crc kubenswrapper[4907]: I0313 14:20:04.879072 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556854-l7l9t"] Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.118357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" event={"ID":"257ee294-461c-4344-9dcb-b44f1497f8bf","Type":"ContainerDied","Data":"4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f"} Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.118394 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f" Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.118495 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556860-ljdj7" Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.120003 4907 generic.go:334] "Generic (PLEG): container finished" podID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerID="311a031e1ff4ef3b74d10161b82c176889b7b82821fc88f2a4edeca5e5657ee5" exitCode=0 Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.120050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wds2z" event={"ID":"b501aea0-66d1-4655-b67f-c6a3bacb7cdb","Type":"ContainerDied","Data":"311a031e1ff4ef3b74d10161b82c176889b7b82821fc88f2a4edeca5e5657ee5"} Mar 13 14:20:05 crc kubenswrapper[4907]: E0313 14:20:05.178254 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod257ee294_461c_4344_9dcb_b44f1497f8bf.slice/crio-4054e86db3967d65bfa7504ca3ded70759d07d9bcb2b6c5f7340940d6a6f850f\": RecentStats: unable to find data in memory cache]" Mar 13 14:20:05 crc kubenswrapper[4907]: I0313 14:20:05.794257 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82df106b-e5b8-4798-942a-999c2ac7a4be" path="/var/lib/kubelet/pods/82df106b-e5b8-4798-942a-999c2ac7a4be/volumes" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.408726 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.432839 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") pod \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.432933 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") pod \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.432982 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") pod \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\" (UID: \"b501aea0-66d1-4655-b67f-c6a3bacb7cdb\") " Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.433257 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "b501aea0-66d1-4655-b67f-c6a3bacb7cdb" (UID: "b501aea0-66d1-4655-b67f-c6a3bacb7cdb"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.439109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm" (OuterVolumeSpecName: "kube-api-access-fdjxm") pod "b501aea0-66d1-4655-b67f-c6a3bacb7cdb" (UID: "b501aea0-66d1-4655-b67f-c6a3bacb7cdb"). InnerVolumeSpecName "kube-api-access-fdjxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.458774 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "b501aea0-66d1-4655-b67f-c6a3bacb7cdb" (UID: "b501aea0-66d1-4655-b67f-c6a3bacb7cdb"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.535080 4907 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-node-mnt\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.535143 4907 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-crc-storage\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:06 crc kubenswrapper[4907]: I0313 14:20:06.535163 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdjxm\" (UniqueName: \"kubernetes.io/projected/b501aea0-66d1-4655-b67f-c6a3bacb7cdb-kube-api-access-fdjxm\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:07 crc kubenswrapper[4907]: I0313 14:20:07.137441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-wds2z" event={"ID":"b501aea0-66d1-4655-b67f-c6a3bacb7cdb","Type":"ContainerDied","Data":"5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a"} Mar 13 14:20:07 crc kubenswrapper[4907]: I0313 14:20:07.137863 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e8decbb8a440a87fe3c9efeb0e3bba54f69864a8dedb5e80d99fcf912b2628a" Mar 13 14:20:07 crc kubenswrapper[4907]: I0313 14:20:07.137552 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-wds2z" Mar 13 14:20:13 crc kubenswrapper[4907]: I0313 14:20:13.048651 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8pphf" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.750858 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s"] Mar 13 14:20:14 crc kubenswrapper[4907]: E0313 14:20:14.751616 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerName="storage" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751643 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerName="storage" Mar 13 14:20:14 crc kubenswrapper[4907]: E0313 14:20:14.751665 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerName="oc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751679 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerName="oc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751951 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" containerName="oc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.751997 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" containerName="storage" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.753477 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.755778 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.763645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s"] Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.854847 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.855113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.855335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.957164 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.957403 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.957510 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.958034 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.958044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:14 crc kubenswrapper[4907]: I0313 14:20:14.979299 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:15 crc kubenswrapper[4907]: I0313 14:20:15.088036 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:15 crc kubenswrapper[4907]: I0313 14:20:15.504733 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s"] Mar 13 14:20:16 crc kubenswrapper[4907]: I0313 14:20:16.191675 4907 generic.go:334] "Generic (PLEG): container finished" podID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerID="8681768c7f655c9c29e81df6bafc91d92b79652dac2656bc6393664d83a817a3" exitCode=0 Mar 13 14:20:16 crc kubenswrapper[4907]: I0313 14:20:16.191759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"8681768c7f655c9c29e81df6bafc91d92b79652dac2656bc6393664d83a817a3"} Mar 13 14:20:16 crc kubenswrapper[4907]: I0313 14:20:16.191961 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerStarted","Data":"9144c72293eb14856491b3ee7a3a07c84f6294a348c2752001d7e79b3abf4e79"} Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.094393 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.095734 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.109079 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.289783 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.290052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.290075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.391534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.391590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.391673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.392175 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.392213 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.410452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"redhat-operators-sftcn\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.433072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:17 crc kubenswrapper[4907]: I0313 14:20:17.658068 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:17 crc kubenswrapper[4907]: W0313 14:20:17.666041 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1885cbd6_ea2e_4d89_b9be_5d8ab9d8d959.slice/crio-32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8 WatchSource:0}: Error finding container 32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8: Status 404 returned error can't find the container with id 32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8 Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.204510 4907 generic.go:334] "Generic (PLEG): container finished" podID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" exitCode=0 Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.204624 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee"} Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.204969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerStarted","Data":"32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8"} Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.206925 4907 generic.go:334] "Generic (PLEG): container finished" podID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerID="b1907d331967d1a964d1cfc7f63d248b5cfe9d2b0d45e04047cb097b8077faf3" exitCode=0 Mar 13 14:20:18 crc kubenswrapper[4907]: I0313 14:20:18.206948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"b1907d331967d1a964d1cfc7f63d248b5cfe9d2b0d45e04047cb097b8077faf3"} Mar 13 14:20:19 crc kubenswrapper[4907]: I0313 14:20:19.217560 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerStarted","Data":"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc"} Mar 13 14:20:19 crc kubenswrapper[4907]: I0313 14:20:19.222458 4907 generic.go:334] "Generic (PLEG): container finished" podID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerID="2e38829b5a9579bee5bd078fe0fab962d0e0f2f365a6b75024fe01a2cd8d9e8a" exitCode=0 Mar 13 14:20:19 crc kubenswrapper[4907]: I0313 14:20:19.222584 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"2e38829b5a9579bee5bd078fe0fab962d0e0f2f365a6b75024fe01a2cd8d9e8a"} Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.231556 4907 generic.go:334] "Generic (PLEG): container finished" podID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" exitCode=0 Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.231663 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc"} Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.629352 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.832478 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") pod \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.832711 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") pod \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.832803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") pod \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\" (UID: \"a387235c-ef5d-47cc-8dff-8af5e8b5690f\") " Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.833569 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle" (OuterVolumeSpecName: "bundle") pod "a387235c-ef5d-47cc-8dff-8af5e8b5690f" (UID: "a387235c-ef5d-47cc-8dff-8af5e8b5690f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.839641 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4" (OuterVolumeSpecName: "kube-api-access-qrhj4") pod "a387235c-ef5d-47cc-8dff-8af5e8b5690f" (UID: "a387235c-ef5d-47cc-8dff-8af5e8b5690f"). InnerVolumeSpecName "kube-api-access-qrhj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.852535 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util" (OuterVolumeSpecName: "util") pod "a387235c-ef5d-47cc-8dff-8af5e8b5690f" (UID: "a387235c-ef5d-47cc-8dff-8af5e8b5690f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.934501 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrhj4\" (UniqueName: \"kubernetes.io/projected/a387235c-ef5d-47cc-8dff-8af5e8b5690f-kube-api-access-qrhj4\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.934537 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:20 crc kubenswrapper[4907]: I0313 14:20:20.934550 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a387235c-ef5d-47cc-8dff-8af5e8b5690f-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.239661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerStarted","Data":"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356"} Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.243545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" event={"ID":"a387235c-ef5d-47cc-8dff-8af5e8b5690f","Type":"ContainerDied","Data":"9144c72293eb14856491b3ee7a3a07c84f6294a348c2752001d7e79b3abf4e79"} Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.243597 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9144c72293eb14856491b3ee7a3a07c84f6294a348c2752001d7e79b3abf4e79" Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.243609 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s" Mar 13 14:20:21 crc kubenswrapper[4907]: I0313 14:20:21.676047 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sftcn" podStartSLOduration=2.083378881 podStartE2EDuration="4.676028288s" podCreationTimestamp="2026-03-13 14:20:17 +0000 UTC" firstStartedPulling="2026-03-13 14:20:18.20620721 +0000 UTC m=+917.105994899" lastFinishedPulling="2026-03-13 14:20:20.798856607 +0000 UTC m=+919.698644306" observedRunningTime="2026-03-13 14:20:21.258513739 +0000 UTC m=+920.158301448" watchObservedRunningTime="2026-03-13 14:20:21.676028288 +0000 UTC m=+920.575815977" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.994393 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh"] Mar 13 14:20:24 crc kubenswrapper[4907]: E0313 14:20:24.995267 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="extract" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995289 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="extract" Mar 13 14:20:24 crc kubenswrapper[4907]: E0313 14:20:24.995316 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="util" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995324 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="util" Mar 13 14:20:24 crc kubenswrapper[4907]: E0313 14:20:24.995344 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="pull" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995353 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="pull" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.995476 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a387235c-ef5d-47cc-8dff-8af5e8b5690f" containerName="extract" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.996060 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.997963 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.998332 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4xlqd" Mar 13 14:20:24 crc kubenswrapper[4907]: I0313 14:20:24.998382 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.031417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh"] Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.188192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjjcp\" (UniqueName: \"kubernetes.io/projected/443989a2-3471-4c69-a460-c441d90064fe-kube-api-access-vjjcp\") pod \"nmstate-operator-796d4cfff4-jrdzh\" (UID: \"443989a2-3471-4c69-a460-c441d90064fe\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.289082 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjjcp\" (UniqueName: \"kubernetes.io/projected/443989a2-3471-4c69-a460-c441d90064fe-kube-api-access-vjjcp\") pod \"nmstate-operator-796d4cfff4-jrdzh\" (UID: \"443989a2-3471-4c69-a460-c441d90064fe\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.308579 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjjcp\" (UniqueName: \"kubernetes.io/projected/443989a2-3471-4c69-a460-c441d90064fe-kube-api-access-vjjcp\") pod \"nmstate-operator-796d4cfff4-jrdzh\" (UID: \"443989a2-3471-4c69-a460-c441d90064fe\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.366593 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" Mar 13 14:20:25 crc kubenswrapper[4907]: I0313 14:20:25.602756 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh"] Mar 13 14:20:26 crc kubenswrapper[4907]: I0313 14:20:26.269896 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" event={"ID":"443989a2-3471-4c69-a460-c441d90064fe","Type":"ContainerStarted","Data":"8a754bbd1938636e6c7ebb1e6c1879692f0daf6d651e0cbf15f733a870f4a158"} Mar 13 14:20:27 crc kubenswrapper[4907]: I0313 14:20:27.434160 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:27 crc kubenswrapper[4907]: I0313 14:20:27.434210 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:28 crc kubenswrapper[4907]: I0313 14:20:28.484251 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sftcn" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" probeResult="failure" output=< Mar 13 14:20:28 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:20:28 crc kubenswrapper[4907]: > Mar 13 14:20:30 crc kubenswrapper[4907]: I0313 14:20:30.297659 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" event={"ID":"443989a2-3471-4c69-a460-c441d90064fe","Type":"ContainerStarted","Data":"24f35f39833c1827d4c2274e72061a384384a16f92297d22076b13413798882e"} Mar 13 14:20:30 crc kubenswrapper[4907]: I0313 14:20:30.316422 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-jrdzh" podStartSLOduration=2.273083783 podStartE2EDuration="6.316399174s" podCreationTimestamp="2026-03-13 14:20:24 +0000 UTC" firstStartedPulling="2026-03-13 14:20:25.62311446 +0000 UTC m=+924.522902149" lastFinishedPulling="2026-03-13 14:20:29.666429851 +0000 UTC m=+928.566217540" observedRunningTime="2026-03-13 14:20:30.313464662 +0000 UTC m=+929.213252351" watchObservedRunningTime="2026-03-13 14:20:30.316399174 +0000 UTC m=+929.216186863" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.850873 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.852317 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.854199 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nrmnv" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.866726 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.867604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.874548 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.904265 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.913075 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.919995 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-4drs7"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.920764 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.979257 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.980089 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.984429 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h"] Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.985440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.985649 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 13 14:20:34 crc kubenswrapper[4907]: I0313 14:20:34.985830 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vxrgq" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.017409 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qplkl\" (UniqueName: \"kubernetes.io/projected/57da19c3-e374-4e40-911a-459c1d9da8f7-kube-api-access-qplkl\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.017460 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6894k\" (UniqueName: \"kubernetes.io/projected/5e672ba1-22be-4fb4-b11c-3aa6a2b69c81-kube-api-access-6894k\") pod \"nmstate-metrics-9b8c8685d-fnz7h\" (UID: \"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.017737 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/57da19c3-e374-4e40-911a-459c1d9da8f7-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119458 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qplkl\" (UniqueName: \"kubernetes.io/projected/57da19c3-e374-4e40-911a-459c1d9da8f7-kube-api-access-qplkl\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119594 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs2zt\" (UniqueName: \"kubernetes.io/projected/46a0badb-4ddd-4f72-a6ae-c415a5485062-kube-api-access-zs2zt\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119654 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6894k\" (UniqueName: \"kubernetes.io/projected/5e672ba1-22be-4fb4-b11c-3aa6a2b69c81-kube-api-access-6894k\") pod \"nmstate-metrics-9b8c8685d-fnz7h\" (UID: \"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119727 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-nmstate-lock\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119828 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/46a0badb-4ddd-4f72-a6ae-c415a5485062-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119910 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m246\" (UniqueName: \"kubernetes.io/projected/f945fc96-0cb3-469c-ac53-8a37e0b01f51-kube-api-access-4m246\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.119968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/57da19c3-e374-4e40-911a-459c1d9da8f7-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.120021 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-ovs-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.120141 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-dbus-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.120184 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.137904 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/57da19c3-e374-4e40-911a-459c1d9da8f7-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.144161 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qplkl\" (UniqueName: \"kubernetes.io/projected/57da19c3-e374-4e40-911a-459c1d9da8f7-kube-api-access-qplkl\") pod \"nmstate-webhook-5f558f5558-cqjnn\" (UID: \"57da19c3-e374-4e40-911a-459c1d9da8f7\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.147628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6894k\" (UniqueName: \"kubernetes.io/projected/5e672ba1-22be-4fb4-b11c-3aa6a2b69c81-kube-api-access-6894k\") pod \"nmstate-metrics-9b8c8685d-fnz7h\" (UID: \"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.171383 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.199350 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.220117 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-65d688f545-w8jgd"] Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.220902 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs2zt\" (UniqueName: \"kubernetes.io/projected/46a0badb-4ddd-4f72-a6ae-c415a5485062-kube-api-access-zs2zt\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221160 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-nmstate-lock\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221195 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/46a0badb-4ddd-4f72-a6ae-c415a5485062-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m246\" (UniqueName: \"kubernetes.io/projected/f945fc96-0cb3-469c-ac53-8a37e0b01f51-kube-api-access-4m246\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-ovs-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221342 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-nmstate-lock\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.221849 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-ovs-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.222705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/46a0badb-4ddd-4f72-a6ae-c415a5485062-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.222765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-dbus-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.222792 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: E0313 14:20:35.222963 4907 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 13 14:20:35 crc kubenswrapper[4907]: E0313 14:20:35.223029 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert podName:46a0badb-4ddd-4f72-a6ae-c415a5485062 nodeName:}" failed. No retries permitted until 2026-03-13 14:20:35.723010738 +0000 UTC m=+934.622798427 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-wrb6h" (UID: "46a0badb-4ddd-4f72-a6ae-c415a5485062") : secret "plugin-serving-cert" not found Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.223096 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f945fc96-0cb3-469c-ac53-8a37e0b01f51-dbus-socket\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.238324 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-65d688f545-w8jgd"] Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.248091 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs2zt\" (UniqueName: \"kubernetes.io/projected/46a0badb-4ddd-4f72-a6ae-c415a5485062-kube-api-access-zs2zt\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.248093 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m246\" (UniqueName: \"kubernetes.io/projected/f945fc96-0cb3-469c-ac53-8a37e0b01f51-kube-api-access-4m246\") pod \"nmstate-handler-4drs7\" (UID: \"f945fc96-0cb3-469c-ac53-8a37e0b01f51\") " pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.325247 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-oauth-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326185 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-oauth-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326231 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-trusted-ca-bundle\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv7v8\" (UniqueName: \"kubernetes.io/projected/25a755e0-d2c0-414f-8a68-aa0c9c28079c-kube-api-access-wv7v8\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-service-ca\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326374 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.326392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427615 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv7v8\" (UniqueName: \"kubernetes.io/projected/25a755e0-d2c0-414f-8a68-aa0c9c28079c-kube-api-access-wv7v8\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427654 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-service-ca\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427743 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-oauth-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427766 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-oauth-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.427783 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-trusted-ca-bundle\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.428778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-trusted-ca-bundle\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.428894 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-oauth-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.430223 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.431515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/25a755e0-d2c0-414f-8a68-aa0c9c28079c-service-ca\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.433807 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-oauth-config\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.434526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/25a755e0-d2c0-414f-8a68-aa0c9c28079c-console-serving-cert\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.444222 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv7v8\" (UniqueName: \"kubernetes.io/projected/25a755e0-d2c0-414f-8a68-aa0c9c28079c-kube-api-access-wv7v8\") pod \"console-65d688f545-w8jgd\" (UID: \"25a755e0-d2c0-414f-8a68-aa0c9c28079c\") " pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.456008 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn"] Mar 13 14:20:35 crc kubenswrapper[4907]: W0313 14:20:35.457796 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57da19c3_e374_4e40_911a_459c1d9da8f7.slice/crio-9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6 WatchSource:0}: Error finding container 9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6: Status 404 returned error can't find the container with id 9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6 Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.537595 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:35 crc kubenswrapper[4907]: W0313 14:20:35.564512 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf945fc96_0cb3_469c_ac53_8a37e0b01f51.slice/crio-2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b WatchSource:0}: Error finding container 2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b: Status 404 returned error can't find the container with id 2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.601033 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.630688 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h"] Mar 13 14:20:35 crc kubenswrapper[4907]: W0313 14:20:35.636292 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e672ba1_22be_4fb4_b11c_3aa6a2b69c81.slice/crio-d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e WatchSource:0}: Error finding container d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e: Status 404 returned error can't find the container with id d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.730653 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.735626 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/46a0badb-4ddd-4f72-a6ae-c415a5485062-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-wrb6h\" (UID: \"46a0badb-4ddd-4f72-a6ae-c415a5485062\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.800963 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-65d688f545-w8jgd"] Mar 13 14:20:35 crc kubenswrapper[4907]: I0313 14:20:35.897502 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.106841 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h"] Mar 13 14:20:36 crc kubenswrapper[4907]: W0313 14:20:36.111846 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46a0badb_4ddd_4f72_a6ae_c415a5485062.slice/crio-2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7 WatchSource:0}: Error finding container 2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7: Status 404 returned error can't find the container with id 2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7 Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.362734 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" event={"ID":"57da19c3-e374-4e40-911a-459c1d9da8f7","Type":"ContainerStarted","Data":"9046362e8ce8a6b9344f48e54f7ce1f08c91443317b3ba413fbbc123ec101bc6"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.364200 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" event={"ID":"46a0badb-4ddd-4f72-a6ae-c415a5485062","Type":"ContainerStarted","Data":"2b51f1fbe6aaca4e97833e4f8fca8a3a1e0b6f3117495a6860581c66c8562ff7"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.365954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" event={"ID":"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81","Type":"ContainerStarted","Data":"d63fd92a574dd16432e8cc6b1570153bb6e2b7cb09e0e9c81696e31884132f2e"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.368129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-65d688f545-w8jgd" event={"ID":"25a755e0-d2c0-414f-8a68-aa0c9c28079c","Type":"ContainerStarted","Data":"6a7d64edf00f3d3262dc2c6224febcf7d60e0f32623e61efc8de67a598cf7646"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.368161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-65d688f545-w8jgd" event={"ID":"25a755e0-d2c0-414f-8a68-aa0c9c28079c","Type":"ContainerStarted","Data":"1af2fc91e62a0ba1fbf071d2fffc6640ff0e146852d1684462d2d89243a2698a"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.369486 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4drs7" event={"ID":"f945fc96-0cb3-469c-ac53-8a37e0b01f51","Type":"ContainerStarted","Data":"2dff306914b4121425b0b92bc2102810f2dd230dbcd3319f0293820db567826b"} Mar 13 14:20:36 crc kubenswrapper[4907]: I0313 14:20:36.389721 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-65d688f545-w8jgd" podStartSLOduration=1.389702041 podStartE2EDuration="1.389702041s" podCreationTimestamp="2026-03-13 14:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:20:36.386376939 +0000 UTC m=+935.286164618" watchObservedRunningTime="2026-03-13 14:20:36.389702041 +0000 UTC m=+935.289489750" Mar 13 14:20:37 crc kubenswrapper[4907]: I0313 14:20:37.474212 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:37 crc kubenswrapper[4907]: I0313 14:20:37.523290 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:37 crc kubenswrapper[4907]: I0313 14:20:37.703767 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.383455 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4drs7" event={"ID":"f945fc96-0cb3-469c-ac53-8a37e0b01f51","Type":"ContainerStarted","Data":"b63fe2e822839caaa6054f85481e03baef52d46da7bbf472cfdd35feb6cc2ae7"} Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.383791 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.389615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" event={"ID":"57da19c3-e374-4e40-911a-459c1d9da8f7","Type":"ContainerStarted","Data":"5b7751fd2f4a05db76d4f555271c5c0b9ab1b77200a9a6a2dbf746f37e7d52a9"} Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.390236 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.391695 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" event={"ID":"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81","Type":"ContainerStarted","Data":"0733092feefb0ff9118e05d8d0be492c02cc5f9c3b863e4cb7897ca1e6c58c29"} Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.402409 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-4drs7" podStartSLOduration=2.022649089 podStartE2EDuration="4.402384158s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:35.567848402 +0000 UTC m=+934.467636121" lastFinishedPulling="2026-03-13 14:20:37.947583461 +0000 UTC m=+936.847371190" observedRunningTime="2026-03-13 14:20:38.39745281 +0000 UTC m=+937.297240499" watchObservedRunningTime="2026-03-13 14:20:38.402384158 +0000 UTC m=+937.302171867" Mar 13 14:20:38 crc kubenswrapper[4907]: I0313 14:20:38.419493 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" podStartSLOduration=1.907904863 podStartE2EDuration="4.419473883s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:35.459468744 +0000 UTC m=+934.359256433" lastFinishedPulling="2026-03-13 14:20:37.971037764 +0000 UTC m=+936.870825453" observedRunningTime="2026-03-13 14:20:38.41324132 +0000 UTC m=+937.313029009" watchObservedRunningTime="2026-03-13 14:20:38.419473883 +0000 UTC m=+937.319261572" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.409578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" event={"ID":"46a0badb-4ddd-4f72-a6ae-c415a5485062","Type":"ContainerStarted","Data":"0c762005459774c9f2cf4e2c09c71c5c6d8a5001c6e4b196411b93ad10ff0ce7"} Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.410233 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sftcn" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" containerID="cri-o://a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" gracePeriod=2 Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.428247 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-wrb6h" podStartSLOduration=2.715792122 podStartE2EDuration="5.428202737s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:36.113839328 +0000 UTC m=+935.013627017" lastFinishedPulling="2026-03-13 14:20:38.826249933 +0000 UTC m=+937.726037632" observedRunningTime="2026-03-13 14:20:39.423488236 +0000 UTC m=+938.323275925" watchObservedRunningTime="2026-03-13 14:20:39.428202737 +0000 UTC m=+938.327990426" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.745810 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.788802 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") pod \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.793928 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9" (OuterVolumeSpecName: "kube-api-access-zp4z9") pod "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" (UID: "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959"). InnerVolumeSpecName "kube-api-access-zp4z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.889683 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") pod \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.889741 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") pod \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\" (UID: \"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959\") " Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.889956 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp4z9\" (UniqueName: \"kubernetes.io/projected/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-kube-api-access-zp4z9\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.890642 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities" (OuterVolumeSpecName: "utilities") pod "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" (UID: "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:39 crc kubenswrapper[4907]: I0313 14:20:39.991258 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.018820 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" (UID: "1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.092199 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.417434 4907 generic.go:334] "Generic (PLEG): container finished" podID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" exitCode=0 Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.418562 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sftcn" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.429952 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356"} Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.430018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sftcn" event={"ID":"1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959","Type":"ContainerDied","Data":"32ac1d4a8e7dc17b286bf397735fbaec81ae96dd50b7a9ab89b1fd768ac31fd8"} Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.430048 4907 scope.go:117] "RemoveContainer" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.457928 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.465251 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sftcn"] Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.467760 4907 scope.go:117] "RemoveContainer" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.516771 4907 scope.go:117] "RemoveContainer" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.530903 4907 scope.go:117] "RemoveContainer" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" Mar 13 14:20:40 crc kubenswrapper[4907]: E0313 14:20:40.531259 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356\": container with ID starting with a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356 not found: ID does not exist" containerID="a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.531305 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356"} err="failed to get container status \"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356\": rpc error: code = NotFound desc = could not find container \"a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356\": container with ID starting with a9ac0599c640beb63d63a6aae7a564ac63e2bcc0d12ec69b4aa940c25b9e8356 not found: ID does not exist" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.531335 4907 scope.go:117] "RemoveContainer" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" Mar 13 14:20:40 crc kubenswrapper[4907]: E0313 14:20:40.531731 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc\": container with ID starting with df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc not found: ID does not exist" containerID="df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.531762 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc"} err="failed to get container status \"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc\": rpc error: code = NotFound desc = could not find container \"df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc\": container with ID starting with df47c4a79801f3d5add4de851978c2a6b00163c4440a6a765aebda87af0252bc not found: ID does not exist" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.532045 4907 scope.go:117] "RemoveContainer" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" Mar 13 14:20:40 crc kubenswrapper[4907]: E0313 14:20:40.532383 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee\": container with ID starting with c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee not found: ID does not exist" containerID="c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee" Mar 13 14:20:40 crc kubenswrapper[4907]: I0313 14:20:40.532414 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee"} err="failed to get container status \"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee\": rpc error: code = NotFound desc = could not find container \"c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee\": container with ID starting with c6fc8c78c0d2400b40d785366f4fdd78a06e3d1fb3c60436c40f635cf1fa15ee not found: ID does not exist" Mar 13 14:20:41 crc kubenswrapper[4907]: I0313 14:20:41.431161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" event={"ID":"5e672ba1-22be-4fb4-b11c-3aa6a2b69c81","Type":"ContainerStarted","Data":"f40113f178ce01366186258c102e3799aab0ed2e3849d2235a8b7cb62f5bc4a0"} Mar 13 14:20:41 crc kubenswrapper[4907]: I0313 14:20:41.461993 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-fnz7h" podStartSLOduration=2.570348044 podStartE2EDuration="7.461965831s" podCreationTimestamp="2026-03-13 14:20:34 +0000 UTC" firstStartedPulling="2026-03-13 14:20:35.641080592 +0000 UTC m=+934.540868281" lastFinishedPulling="2026-03-13 14:20:40.532698379 +0000 UTC m=+939.432486068" observedRunningTime="2026-03-13 14:20:41.456715044 +0000 UTC m=+940.356502743" watchObservedRunningTime="2026-03-13 14:20:41.461965831 +0000 UTC m=+940.361753570" Mar 13 14:20:41 crc kubenswrapper[4907]: I0313 14:20:41.797366 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" path="/var/lib/kubelet/pods/1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959/volumes" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.562532 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-4drs7" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.602238 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.602301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:45 crc kubenswrapper[4907]: I0313 14:20:45.606472 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:46 crc kubenswrapper[4907]: I0313 14:20:46.475608 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-65d688f545-w8jgd" Mar 13 14:20:46 crc kubenswrapper[4907]: I0313 14:20:46.517973 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008028 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:20:53 crc kubenswrapper[4907]: E0313 14:20:53.008680 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-utilities" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008703 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-utilities" Mar 13 14:20:53 crc kubenswrapper[4907]: E0313 14:20:53.008724 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008735 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" Mar 13 14:20:53 crc kubenswrapper[4907]: E0313 14:20:53.008750 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-content" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008762 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="extract-content" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.008958 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1885cbd6-ea2e-4d89-b9be-5d8ab9d8d959" containerName="registry-server" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.010077 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.025537 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.164451 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.164525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.164573 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.265493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.265566 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.265612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.266110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.266332 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.289011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"certified-operators-pqj8w\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.348497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:20:53 crc kubenswrapper[4907]: I0313 14:20:53.780526 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:20:53 crc kubenswrapper[4907]: W0313 14:20:53.789681 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb19cbd4_fd73_485d_9b1e_a9ed01fa73ac.slice/crio-a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947 WatchSource:0}: Error finding container a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947: Status 404 returned error can't find the container with id a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947 Mar 13 14:20:54 crc kubenswrapper[4907]: I0313 14:20:54.531477 4907 generic.go:334] "Generic (PLEG): container finished" podID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" exitCode=0 Mar 13 14:20:54 crc kubenswrapper[4907]: I0313 14:20:54.531956 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1"} Mar 13 14:20:54 crc kubenswrapper[4907]: I0313 14:20:54.532012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerStarted","Data":"a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947"} Mar 13 14:20:55 crc kubenswrapper[4907]: I0313 14:20:55.205957 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-cqjnn" Mar 13 14:20:55 crc kubenswrapper[4907]: I0313 14:20:55.545584 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerStarted","Data":"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640"} Mar 13 14:20:56 crc kubenswrapper[4907]: I0313 14:20:56.554465 4907 generic.go:334] "Generic (PLEG): container finished" podID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" exitCode=0 Mar 13 14:20:56 crc kubenswrapper[4907]: I0313 14:20:56.554513 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640"} Mar 13 14:20:57 crc kubenswrapper[4907]: I0313 14:20:57.563072 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerStarted","Data":"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778"} Mar 13 14:20:57 crc kubenswrapper[4907]: I0313 14:20:57.580140 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pqj8w" podStartSLOduration=3.114961532 podStartE2EDuration="5.58011902s" podCreationTimestamp="2026-03-13 14:20:52 +0000 UTC" firstStartedPulling="2026-03-13 14:20:54.536493921 +0000 UTC m=+953.436281610" lastFinishedPulling="2026-03-13 14:20:57.001651409 +0000 UTC m=+955.901439098" observedRunningTime="2026-03-13 14:20:57.578336911 +0000 UTC m=+956.478124600" watchObservedRunningTime="2026-03-13 14:20:57.58011902 +0000 UTC m=+956.479906719" Mar 13 14:21:02 crc kubenswrapper[4907]: I0313 14:21:02.489097 4907 scope.go:117] "RemoveContainer" containerID="2fab8f0b7470e3acd4d60ef3af4057ac0e1b8ec4f3de833c7e05ac6623d0bf47" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.350440 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.350862 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.412685 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.642804 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:03 crc kubenswrapper[4907]: I0313 14:21:03.685065 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:21:05 crc kubenswrapper[4907]: I0313 14:21:05.614410 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pqj8w" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" containerID="cri-o://dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" gracePeriod=2 Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.463489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.584056 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") pod \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.584115 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") pod \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.584159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") pod \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\" (UID: \"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac\") " Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.585206 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities" (OuterVolumeSpecName: "utilities") pod "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" (UID: "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.593976 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft" (OuterVolumeSpecName: "kube-api-access-ch7ft") pod "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" (UID: "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac"). InnerVolumeSpecName "kube-api-access-ch7ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626251 4907 generic.go:334] "Generic (PLEG): container finished" podID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" exitCode=0 Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626315 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pqj8w" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626323 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778"} Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pqj8w" event={"ID":"cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac","Type":"ContainerDied","Data":"a30ed723dfd93724481722c622f269a4f11046d220f97345500ef389c7268947"} Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.626673 4907 scope.go:117] "RemoveContainer" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.646246 4907 scope.go:117] "RemoveContainer" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.662916 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" (UID: "cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.664848 4907 scope.go:117] "RemoveContainer" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.686560 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.686585 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ch7ft\" (UniqueName: \"kubernetes.io/projected/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-kube-api-access-ch7ft\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.686595 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.696682 4907 scope.go:117] "RemoveContainer" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" Mar 13 14:21:06 crc kubenswrapper[4907]: E0313 14:21:06.697114 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778\": container with ID starting with dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778 not found: ID does not exist" containerID="dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697205 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778"} err="failed to get container status \"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778\": rpc error: code = NotFound desc = could not find container \"dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778\": container with ID starting with dac05b5102dfe562f7f72b8cbc866e0b01dabd3fb6d1bf5ab854368d949c6778 not found: ID does not exist" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697239 4907 scope.go:117] "RemoveContainer" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" Mar 13 14:21:06 crc kubenswrapper[4907]: E0313 14:21:06.697656 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640\": container with ID starting with a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640 not found: ID does not exist" containerID="a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697694 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640"} err="failed to get container status \"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640\": rpc error: code = NotFound desc = could not find container \"a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640\": container with ID starting with a8fc5a1c00d8ab234f0ce968e9fc0a3754194014ed118abc584a840f27aef640 not found: ID does not exist" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.697724 4907 scope.go:117] "RemoveContainer" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" Mar 13 14:21:06 crc kubenswrapper[4907]: E0313 14:21:06.698093 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1\": container with ID starting with 270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1 not found: ID does not exist" containerID="270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.698114 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1"} err="failed to get container status \"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1\": rpc error: code = NotFound desc = could not find container \"270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1\": container with ID starting with 270f522a63d6b7db266e9f432dbe7c6050c412adf666d82c89c463a2eae54bb1 not found: ID does not exist" Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.965625 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:21:06 crc kubenswrapper[4907]: I0313 14:21:06.972228 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pqj8w"] Mar 13 14:21:07 crc kubenswrapper[4907]: I0313 14:21:07.790805 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" path="/var/lib/kubelet/pods/cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac/volumes" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.512674 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl"] Mar 13 14:21:10 crc kubenswrapper[4907]: E0313 14:21:10.513228 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-utilities" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513243 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-utilities" Mar 13 14:21:10 crc kubenswrapper[4907]: E0313 14:21:10.513266 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-content" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513274 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="extract-content" Mar 13 14:21:10 crc kubenswrapper[4907]: E0313 14:21:10.513289 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513300 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.513427 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb19cbd4-fd73-485d-9b1e-a9ed01fa73ac" containerName="registry-server" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.514384 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.516964 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.523040 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl"] Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.534740 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.534841 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.534974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.650613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651221 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651527 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.651724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.670407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:10 crc kubenswrapper[4907]: I0313 14:21:10.843109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.305041 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl"] Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.579746 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-l9xpb" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" containerID="cri-o://27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" gracePeriod=15 Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.660620 4907 generic.go:334] "Generic (PLEG): container finished" podID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerID="30a631b3f50b179ac1b95bb880442cbed61c80cb1509ef7f12866056fb40c103" exitCode=0 Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.660686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"30a631b3f50b179ac1b95bb880442cbed61c80cb1509ef7f12866056fb40c103"} Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.660717 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerStarted","Data":"ee977785926063e5f131d49c4b9de5e11f45ee3e668dd114d5625a14c04dadc5"} Mar 13 14:21:11 crc kubenswrapper[4907]: I0313 14:21:11.667220 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.018199 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-l9xpb_4c8c0266-5452-4abb-92c5-c536df94dd41/console/0.log" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.018271 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172129 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172168 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172211 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172267 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") pod \"4c8c0266-5452-4abb-92c5-c536df94dd41\" (UID: \"4c8c0266-5452-4abb-92c5-c536df94dd41\") " Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172870 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.172905 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.173342 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config" (OuterVolumeSpecName: "console-config") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.173501 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca" (OuterVolumeSpecName: "service-ca") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.179372 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.186581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.191232 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4" (OuterVolumeSpecName: "kube-api-access-vxlx4") pod "4c8c0266-5452-4abb-92c5-c536df94dd41" (UID: "4c8c0266-5452-4abb-92c5-c536df94dd41"). InnerVolumeSpecName "kube-api-access-vxlx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274401 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxlx4\" (UniqueName: \"kubernetes.io/projected/4c8c0266-5452-4abb-92c5-c536df94dd41-kube-api-access-vxlx4\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274622 4907 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274695 4907 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274754 4907 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274806 4907 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274858 4907 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c8c0266-5452-4abb-92c5-c536df94dd41-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.274974 4907 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c8c0266-5452-4abb-92c5-c536df94dd41-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669478 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-l9xpb_4c8c0266-5452-4abb-92c5-c536df94dd41/console/0.log" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669832 4907 generic.go:334] "Generic (PLEG): container finished" podID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" exitCode=2 Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669874 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerDied","Data":"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa"} Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-l9xpb" event={"ID":"4c8c0266-5452-4abb-92c5-c536df94dd41","Type":"ContainerDied","Data":"2ca75bf0ab6ab7b6d03a36b2b78748709fe1e70503c3e45b40703b8a20dadbca"} Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.669980 4907 scope.go:117] "RemoveContainer" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.670130 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-l9xpb" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.702218 4907 scope.go:117] "RemoveContainer" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" Mar 13 14:21:12 crc kubenswrapper[4907]: E0313 14:21:12.702814 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa\": container with ID starting with 27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa not found: ID does not exist" containerID="27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.702858 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa"} err="failed to get container status \"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa\": rpc error: code = NotFound desc = could not find container \"27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa\": container with ID starting with 27a56c8f2bc88df1fc0af00ca4f2cc9f8243a27d536703bf1460c381978493aa not found: ID does not exist" Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.706770 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:21:12 crc kubenswrapper[4907]: I0313 14:21:12.713679 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-l9xpb"] Mar 13 14:21:13 crc kubenswrapper[4907]: I0313 14:21:13.679337 4907 generic.go:334] "Generic (PLEG): container finished" podID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerID="810cdbdfbf983697c54bd5c7dbb827e3963e513b598bbd0dac35273b14be8d25" exitCode=0 Mar 13 14:21:13 crc kubenswrapper[4907]: I0313 14:21:13.679398 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"810cdbdfbf983697c54bd5c7dbb827e3963e513b598bbd0dac35273b14be8d25"} Mar 13 14:21:13 crc kubenswrapper[4907]: I0313 14:21:13.793144 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" path="/var/lib/kubelet/pods/4c8c0266-5452-4abb-92c5-c536df94dd41/volumes" Mar 13 14:21:14 crc kubenswrapper[4907]: I0313 14:21:14.697559 4907 generic.go:334] "Generic (PLEG): container finished" podID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerID="d6aa27c66a086be71679587fc879df03add237621c2207cd9ce3d8a5d0e7296d" exitCode=0 Mar 13 14:21:14 crc kubenswrapper[4907]: I0313 14:21:14.697644 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"d6aa27c66a086be71679587fc879df03add237621c2207cd9ce3d8a5d0e7296d"} Mar 13 14:21:15 crc kubenswrapper[4907]: I0313 14:21:15.982065 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.134446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") pod \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.134584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") pod \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.134706 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") pod \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\" (UID: \"385226e5-9bb7-4f71-888c-fdd2c3a59b6e\") " Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.137824 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle" (OuterVolumeSpecName: "bundle") pod "385226e5-9bb7-4f71-888c-fdd2c3a59b6e" (UID: "385226e5-9bb7-4f71-888c-fdd2c3a59b6e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.140016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w" (OuterVolumeSpecName: "kube-api-access-4sr9w") pod "385226e5-9bb7-4f71-888c-fdd2c3a59b6e" (UID: "385226e5-9bb7-4f71-888c-fdd2c3a59b6e"). InnerVolumeSpecName "kube-api-access-4sr9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.153475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util" (OuterVolumeSpecName: "util") pod "385226e5-9bb7-4f71-888c-fdd2c3a59b6e" (UID: "385226e5-9bb7-4f71-888c-fdd2c3a59b6e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.236944 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.236993 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sr9w\" (UniqueName: \"kubernetes.io/projected/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-kube-api-access-4sr9w\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.237014 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/385226e5-9bb7-4f71-888c-fdd2c3a59b6e-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.713842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" event={"ID":"385226e5-9bb7-4f71-888c-fdd2c3a59b6e","Type":"ContainerDied","Data":"ee977785926063e5f131d49c4b9de5e11f45ee3e668dd114d5625a14c04dadc5"} Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.713919 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee977785926063e5f131d49c4b9de5e11f45ee3e668dd114d5625a14c04dadc5" Mar 13 14:21:16 crc kubenswrapper[4907]: I0313 14:21:16.714031 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl" Mar 13 14:21:18 crc kubenswrapper[4907]: I0313 14:21:18.041478 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:21:18 crc kubenswrapper[4907]: I0313 14:21:18.041857 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948257 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls"] Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948815 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="util" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="util" Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948840 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="pull" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948845 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="pull" Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948853 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="extract" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948859 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="extract" Mar 13 14:21:25 crc kubenswrapper[4907]: E0313 14:21:25.948872 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948878 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948972 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c8c0266-5452-4abb-92c5-c536df94dd41" containerName="console" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.948986 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="385226e5-9bb7-4f71-888c-fdd2c3a59b6e" containerName="extract" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.949313 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.952932 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zzb7m" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.953159 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.956717 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.956937 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.963356 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 13 14:21:25 crc kubenswrapper[4907]: I0313 14:21:25.969383 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.094143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88mvb\" (UniqueName: \"kubernetes.io/projected/1c19fee6-f4bd-4488-acd9-23a4835edd83-kube-api-access-88mvb\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.094199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-apiservice-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.094248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-webhook-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.167226 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.167852 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.169947 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.170222 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.170343 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vtpkb" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.192490 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.194771 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-webhook-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.194831 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88mvb\" (UniqueName: \"kubernetes.io/projected/1c19fee6-f4bd-4488-acd9-23a4835edd83-kube-api-access-88mvb\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.194865 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-apiservice-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.200437 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-apiservice-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.218985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88mvb\" (UniqueName: \"kubernetes.io/projected/1c19fee6-f4bd-4488-acd9-23a4835edd83-kube-api-access-88mvb\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.221685 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1c19fee6-f4bd-4488-acd9-23a4835edd83-webhook-cert\") pod \"metallb-operator-controller-manager-cf9f9b596-sd4ls\" (UID: \"1c19fee6-f4bd-4488-acd9-23a4835edd83\") " pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.267326 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.295737 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.295801 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-webhook-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.295858 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njlcd\" (UniqueName: \"kubernetes.io/projected/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-kube-api-access-njlcd\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.397046 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njlcd\" (UniqueName: \"kubernetes.io/projected/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-kube-api-access-njlcd\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.397401 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.397430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-webhook-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.400745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-webhook-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.403448 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-apiservice-cert\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.416199 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njlcd\" (UniqueName: \"kubernetes.io/projected/211e8a4b-dd3b-4669-b426-f0ac30bb3ca2-kube-api-access-njlcd\") pod \"metallb-operator-webhook-server-bf6b45999-8mfb6\" (UID: \"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2\") " pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.480528 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.690376 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6"] Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.746000 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls"] Mar 13 14:21:26 crc kubenswrapper[4907]: W0313 14:21:26.752405 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c19fee6_f4bd_4488_acd9_23a4835edd83.slice/crio-5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0 WatchSource:0}: Error finding container 5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0: Status 404 returned error can't find the container with id 5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0 Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.767243 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" event={"ID":"1c19fee6-f4bd-4488-acd9-23a4835edd83","Type":"ContainerStarted","Data":"5a45331b0fbfdd4da374f224485bc37c8953dfddabdb352dc960db3fe447c1a0"} Mar 13 14:21:26 crc kubenswrapper[4907]: I0313 14:21:26.768321 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" event={"ID":"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2","Type":"ContainerStarted","Data":"01fd754435af4816d44cee7c7358e62423e9beb748c1d066d0c36016623e5098"} Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.806086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" event={"ID":"1c19fee6-f4bd-4488-acd9-23a4835edd83","Type":"ContainerStarted","Data":"aaf611434f75db6dfcd0b5f6ac99f50e8b3151fa4b3b2d139729f6b42c5a5d75"} Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.806690 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.807604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" event={"ID":"211e8a4b-dd3b-4669-b426-f0ac30bb3ca2","Type":"ContainerStarted","Data":"25dafd6957026f609842bb395e5b42b5f4f7b10c5a859c6dae3f8f705a216d24"} Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.807760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.857253 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" podStartSLOduration=4.119773106 podStartE2EDuration="6.857234375s" podCreationTimestamp="2026-03-13 14:21:25 +0000 UTC" firstStartedPulling="2026-03-13 14:21:26.755629383 +0000 UTC m=+985.655417082" lastFinishedPulling="2026-03-13 14:21:29.493090662 +0000 UTC m=+988.392878351" observedRunningTime="2026-03-13 14:21:31.853025979 +0000 UTC m=+990.752813688" watchObservedRunningTime="2026-03-13 14:21:31.857234375 +0000 UTC m=+990.757022084" Mar 13 14:21:31 crc kubenswrapper[4907]: I0313 14:21:31.873353 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" podStartSLOduration=1.414381437 podStartE2EDuration="5.87333694s" podCreationTimestamp="2026-03-13 14:21:26 +0000 UTC" firstStartedPulling="2026-03-13 14:21:26.697105605 +0000 UTC m=+985.596893294" lastFinishedPulling="2026-03-13 14:21:31.156061108 +0000 UTC m=+990.055848797" observedRunningTime="2026-03-13 14:21:31.872435025 +0000 UTC m=+990.772222704" watchObservedRunningTime="2026-03-13 14:21:31.87333694 +0000 UTC m=+990.773124629" Mar 13 14:21:46 crc kubenswrapper[4907]: I0313 14:21:46.504960 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-bf6b45999-8mfb6" Mar 13 14:21:48 crc kubenswrapper[4907]: I0313 14:21:48.042007 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:21:48 crc kubenswrapper[4907]: I0313 14:21:48.042604 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.136015 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.137508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.141257 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.141809 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.143278 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.147197 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.250543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"auto-csr-approver-29556862-sczxs\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.352383 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"auto-csr-approver-29556862-sczxs\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.389965 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"auto-csr-approver-29556862-sczxs\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.467851 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.672562 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:22:00 crc kubenswrapper[4907]: W0313 14:22:00.675873 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72e27ed7_be4f_4d90_ac02_fa7d64585d6d.slice/crio-caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da WatchSource:0}: Error finding container caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da: Status 404 returned error can't find the container with id caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da Mar 13 14:22:00 crc kubenswrapper[4907]: I0313 14:22:00.996866 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556862-sczxs" event={"ID":"72e27ed7-be4f-4d90-ac02-fa7d64585d6d","Type":"ContainerStarted","Data":"caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da"} Mar 13 14:22:03 crc kubenswrapper[4907]: I0313 14:22:03.023330 4907 generic.go:334] "Generic (PLEG): container finished" podID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerID="ff874f38ed077cec3c5651ce4ba0b1fcf1996e7c127b7d6b128f6c5552e26bed" exitCode=0 Mar 13 14:22:03 crc kubenswrapper[4907]: I0313 14:22:03.023387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556862-sczxs" event={"ID":"72e27ed7-be4f-4d90-ac02-fa7d64585d6d","Type":"ContainerDied","Data":"ff874f38ed077cec3c5651ce4ba0b1fcf1996e7c127b7d6b128f6c5552e26bed"} Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.344014 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.505213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") pod \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\" (UID: \"72e27ed7-be4f-4d90-ac02-fa7d64585d6d\") " Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.515130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc" (OuterVolumeSpecName: "kube-api-access-s8bmc") pod "72e27ed7-be4f-4d90-ac02-fa7d64585d6d" (UID: "72e27ed7-be4f-4d90-ac02-fa7d64585d6d"). InnerVolumeSpecName "kube-api-access-s8bmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:22:04 crc kubenswrapper[4907]: I0313 14:22:04.607418 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8bmc\" (UniqueName: \"kubernetes.io/projected/72e27ed7-be4f-4d90-ac02-fa7d64585d6d-kube-api-access-s8bmc\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.035868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556862-sczxs" event={"ID":"72e27ed7-be4f-4d90-ac02-fa7d64585d6d","Type":"ContainerDied","Data":"caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da"} Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.035927 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caaccae846433276178ae1d1425b6650200856ae6f5f00b62a4c34787d3ca1da" Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.035942 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556862-sczxs" Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.417211 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.421016 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556856-8grxd"] Mar 13 14:22:05 crc kubenswrapper[4907]: I0313 14:22:05.792034 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1604551e-a53d-483c-a03e-3c2159768001" path="/var/lib/kubelet/pods/1604551e-a53d-483c-a03e-3c2159768001/volumes" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.269414 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-cf9f9b596-sd4ls" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.895849 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9"] Mar 13 14:22:06 crc kubenswrapper[4907]: E0313 14:22:06.896141 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerName="oc" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.896154 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerName="oc" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.896264 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" containerName="oc" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.896669 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.899510 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6r77v"] Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.901685 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909020 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909221 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-472l6" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909775 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.909905 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.919665 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9"] Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.991770 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-fkplv"] Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.993201 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fkplv" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.997210 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 13 14:22:06 crc kubenswrapper[4907]: I0313 14:22:06.997907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:06.998683 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dsr7f" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.001058 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.006627 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-2g4kt"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.007714 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.009470 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.018938 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-2g4kt"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.038981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-sockets\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-conf\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039068 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-reloader\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l8tg\" (UniqueName: \"kubernetes.io/projected/f5203ff6-d473-484d-ab8f-414e9ca804d4-kube-api-access-7l8tg\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039110 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/21498ef9-03ec-44de-ade5-cc56f504499b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics-certs\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039157 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-startup\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039175 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x42bp\" (UniqueName: \"kubernetes.io/projected/21498ef9-03ec-44de-ade5-cc56f504499b-kube-api-access-x42bp\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.039203 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-conf\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140585 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-metrics-certs\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140625 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-reloader\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140661 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l8tg\" (UniqueName: \"kubernetes.io/projected/f5203ff6-d473-484d-ab8f-414e9ca804d4-kube-api-access-7l8tg\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140683 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/21498ef9-03ec-44de-ade5-cc56f504499b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.140963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-conf\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-reloader\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141527 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics-certs\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5564ca73-f820-49a4-affa-fd718d1ad7af-metallb-excludel2\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141609 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twrdg\" (UniqueName: \"kubernetes.io/projected/255dfb62-fe3e-440f-82ed-1ff604426a9b-kube-api-access-twrdg\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141671 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-startup\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141694 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhdph\" (UniqueName: \"kubernetes.io/projected/5564ca73-f820-49a4-affa-fd718d1ad7af-kube-api-access-zhdph\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141727 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x42bp\" (UniqueName: \"kubernetes.io/projected/21498ef9-03ec-44de-ade5-cc56f504499b-kube-api-access-x42bp\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-sockets\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141818 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.141839 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-cert\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.142295 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.142389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-startup\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.142505 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/f5203ff6-d473-484d-ab8f-414e9ca804d4-frr-sockets\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.157163 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/21498ef9-03ec-44de-ade5-cc56f504499b-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.158228 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5203ff6-d473-484d-ab8f-414e9ca804d4-metrics-certs\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.162265 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l8tg\" (UniqueName: \"kubernetes.io/projected/f5203ff6-d473-484d-ab8f-414e9ca804d4-kube-api-access-7l8tg\") pod \"frr-k8s-6r77v\" (UID: \"f5203ff6-d473-484d-ab8f-414e9ca804d4\") " pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.162335 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x42bp\" (UniqueName: \"kubernetes.io/projected/21498ef9-03ec-44de-ade5-cc56f504499b-kube-api-access-x42bp\") pod \"frr-k8s-webhook-server-bcc4b6f68-jlsw9\" (UID: \"21498ef9-03ec-44de-ade5-cc56f504499b\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.234227 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243319 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twrdg\" (UniqueName: \"kubernetes.io/projected/255dfb62-fe3e-440f-82ed-1ff604426a9b-kube-api-access-twrdg\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5564ca73-f820-49a4-affa-fd718d1ad7af-metallb-excludel2\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243422 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhdph\" (UniqueName: \"kubernetes.io/projected/5564ca73-f820-49a4-affa-fd718d1ad7af-kube-api-access-zhdph\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243512 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-cert\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.243566 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-metrics-certs\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.243886 4907 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.243974 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist podName:5564ca73-f820-49a4-affa-fd718d1ad7af nodeName:}" failed. No retries permitted until 2026-03-13 14:22:07.743954407 +0000 UTC m=+1026.643742096 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist") pod "speaker-fkplv" (UID: "5564ca73-f820-49a4-affa-fd718d1ad7af") : secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.244939 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.245266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/5564ca73-f820-49a4-affa-fd718d1ad7af-metallb-excludel2\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.246486 4907 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.246639 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs podName:255dfb62-fe3e-440f-82ed-1ff604426a9b nodeName:}" failed. No retries permitted until 2026-03-13 14:22:07.74662049 +0000 UTC m=+1026.646408189 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs") pod "controller-7bb4cc7c98-2g4kt" (UID: "255dfb62-fe3e-440f-82ed-1ff604426a9b") : secret "controller-certs-secret" not found Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.249407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-metrics-certs\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.275205 4907 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.285616 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twrdg\" (UniqueName: \"kubernetes.io/projected/255dfb62-fe3e-440f-82ed-1ff604426a9b-kube-api-access-twrdg\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.296555 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-cert\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.322453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhdph\" (UniqueName: \"kubernetes.io/projected/5564ca73-f820-49a4-affa-fd718d1ad7af-kube-api-access-zhdph\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.515839 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.750157 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.750259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.750370 4907 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: E0313 14:22:07.750431 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist podName:5564ca73-f820-49a4-affa-fd718d1ad7af nodeName:}" failed. No retries permitted until 2026-03-13 14:22:08.750416612 +0000 UTC m=+1027.650204301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist") pod "speaker-fkplv" (UID: "5564ca73-f820-49a4-affa-fd718d1ad7af") : secret "metallb-memberlist" not found Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.756532 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/255dfb62-fe3e-440f-82ed-1ff604426a9b-metrics-certs\") pod \"controller-7bb4cc7c98-2g4kt\" (UID: \"255dfb62-fe3e-440f-82ed-1ff604426a9b\") " pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.862864 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.864583 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.878273 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.927581 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.953242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.953305 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:07 crc kubenswrapper[4907]: I0313 14:22:07.953330 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055629 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.055683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.065871 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" event={"ID":"21498ef9-03ec-44de-ade5-cc56f504499b","Type":"ContainerStarted","Data":"e3c1b667e890611fca7f8facb65f2255ec3d081d5d6caadf21a54230dd63cae4"} Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.068147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"56f5bcfba3b8a1ee28a14907c426f784464a16911d4abf852784c13a3c742e3f"} Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.076778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"community-operators-d4jn9\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.161616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-2g4kt"] Mar 13 14:22:08 crc kubenswrapper[4907]: W0313 14:22:08.170954 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod255dfb62_fe3e_440f_82ed_1ff604426a9b.slice/crio-9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd WatchSource:0}: Error finding container 9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd: Status 404 returned error can't find the container with id 9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.180895 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.527186 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.767784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.775575 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/5564ca73-f820-49a4-affa-fd718d1ad7af-memberlist\") pod \"speaker-fkplv\" (UID: \"5564ca73-f820-49a4-affa-fd718d1ad7af\") " pod="metallb-system/speaker-fkplv" Mar 13 14:22:08 crc kubenswrapper[4907]: I0313 14:22:08.814139 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-fkplv" Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-2g4kt" event={"ID":"255dfb62-fe3e-440f-82ed-1ff604426a9b","Type":"ContainerStarted","Data":"546c6c32d1aa7efc921fd1cd99597c6e7498e802373b3bd72c2dfb820ad5bbad"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077466 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-2g4kt" event={"ID":"255dfb62-fe3e-440f-82ed-1ff604426a9b","Type":"ContainerStarted","Data":"bde2e3680ec5095a34b683394b0a19b7d7cfc0c2a683b3c8d5154da577f8bd79"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.077489 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-2g4kt" event={"ID":"255dfb62-fe3e-440f-82ed-1ff604426a9b","Type":"ContainerStarted","Data":"9bf9ea49769c0e80dc1e8af3b47bb1a01df01b53a39b033db4d913049dbaa5dd"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.080989 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fkplv" event={"ID":"5564ca73-f820-49a4-affa-fd718d1ad7af","Type":"ContainerStarted","Data":"f211dd70f0a3af32a37f299297d692f784b97297cafc168f9d7b88a1639001e6"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.081028 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fkplv" event={"ID":"5564ca73-f820-49a4-affa-fd718d1ad7af","Type":"ContainerStarted","Data":"174f46432910e0a6e62d1e4d69e213487934101226dbd3e79892f20e63afe1b5"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.082552 4907 generic.go:334] "Generic (PLEG): container finished" podID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerID="4cc7ca19806ba6c30546ebc2e565b8d92a5693e0b133afb20dfc043acb245e70" exitCode=0 Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.082599 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"4cc7ca19806ba6c30546ebc2e565b8d92a5693e0b133afb20dfc043acb245e70"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.082624 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerStarted","Data":"9ed471a362cc5821dd762507d4b70cdc1ef3e27ed3258ebfcd19ca5d91f1ce3f"} Mar 13 14:22:09 crc kubenswrapper[4907]: I0313 14:22:09.094287 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-2g4kt" podStartSLOduration=3.0942682 podStartE2EDuration="3.0942682s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:22:09.090772823 +0000 UTC m=+1027.990560532" watchObservedRunningTime="2026-03-13 14:22:09.0942682 +0000 UTC m=+1027.994055889" Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.094171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-fkplv" event={"ID":"5564ca73-f820-49a4-affa-fd718d1ad7af","Type":"ContainerStarted","Data":"6b4f1ac5209d94c3433caa6ad4775427996e7d7ed104c04a12e543aa7f15299f"} Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.094608 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-fkplv" Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.096985 4907 generic.go:334] "Generic (PLEG): container finished" podID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerID="1006ba55273fa47807d756205c115736def78e9d78f5d0bc2e7b68771fa457bf" exitCode=0 Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.097335 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"1006ba55273fa47807d756205c115736def78e9d78f5d0bc2e7b68771fa457bf"} Mar 13 14:22:10 crc kubenswrapper[4907]: I0313 14:22:10.118861 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-fkplv" podStartSLOduration=4.118843463 podStartE2EDuration="4.118843463s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:22:10.115255284 +0000 UTC m=+1029.015042983" watchObservedRunningTime="2026-03-13 14:22:10.118843463 +0000 UTC m=+1029.018631152" Mar 13 14:22:11 crc kubenswrapper[4907]: I0313 14:22:11.109572 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerStarted","Data":"a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a"} Mar 13 14:22:11 crc kubenswrapper[4907]: I0313 14:22:11.129530 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d4jn9" podStartSLOduration=2.679355529 podStartE2EDuration="4.129512544s" podCreationTimestamp="2026-03-13 14:22:07 +0000 UTC" firstStartedPulling="2026-03-13 14:22:09.089807756 +0000 UTC m=+1027.989595445" lastFinishedPulling="2026-03-13 14:22:10.539964781 +0000 UTC m=+1029.439752460" observedRunningTime="2026-03-13 14:22:11.129140663 +0000 UTC m=+1030.028928352" watchObservedRunningTime="2026-03-13 14:22:11.129512544 +0000 UTC m=+1030.029300233" Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.146648 4907 generic.go:334] "Generic (PLEG): container finished" podID="f5203ff6-d473-484d-ab8f-414e9ca804d4" containerID="ef64bf9c95353ba32a6b642cc6315c18c1791829c19b951c20bd6c3207c661ff" exitCode=0 Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.146792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerDied","Data":"ef64bf9c95353ba32a6b642cc6315c18c1791829c19b951c20bd6c3207c661ff"} Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.152846 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" event={"ID":"21498ef9-03ec-44de-ade5-cc56f504499b","Type":"ContainerStarted","Data":"05914453e76dad3d66b07b4cb9ed0eeb79c3f0bd773c8b86deba22bd595b08ad"} Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.153000 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:15 crc kubenswrapper[4907]: I0313 14:22:15.191590 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" podStartSLOduration=1.9440115919999998 podStartE2EDuration="9.191560298s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="2026-03-13 14:22:07.522412181 +0000 UTC m=+1026.422199860" lastFinishedPulling="2026-03-13 14:22:14.769960847 +0000 UTC m=+1033.669748566" observedRunningTime="2026-03-13 14:22:15.184634056 +0000 UTC m=+1034.084421745" watchObservedRunningTime="2026-03-13 14:22:15.191560298 +0000 UTC m=+1034.091347987" Mar 13 14:22:16 crc kubenswrapper[4907]: I0313 14:22:16.162158 4907 generic.go:334] "Generic (PLEG): container finished" podID="f5203ff6-d473-484d-ab8f-414e9ca804d4" containerID="40f2f957691e4f361abb6febbbedd135054fe0358dc3de7631f107f4850803fb" exitCode=0 Mar 13 14:22:16 crc kubenswrapper[4907]: I0313 14:22:16.162227 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerDied","Data":"40f2f957691e4f361abb6febbbedd135054fe0358dc3de7631f107f4850803fb"} Mar 13 14:22:17 crc kubenswrapper[4907]: I0313 14:22:17.170080 4907 generic.go:334] "Generic (PLEG): container finished" podID="f5203ff6-d473-484d-ab8f-414e9ca804d4" containerID="67fa2c802fc206f29dd1e5d6af6be86dda8bb45ad7adda57ab07a6246fe5da5c" exitCode=0 Mar 13 14:22:17 crc kubenswrapper[4907]: I0313 14:22:17.170157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerDied","Data":"67fa2c802fc206f29dd1e5d6af6be86dda8bb45ad7adda57ab07a6246fe5da5c"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.041354 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.041708 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.041752 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.042409 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.042467 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce" gracePeriod=600 Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.181860 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.181927 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.197020 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce" exitCode=0 Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.197102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.197215 4907 scope.go:117] "RemoveContainer" containerID="da83df5fde38cf429de6b21ade144e7c74078db3c329650c84ad8d669e1beb0e" Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204188 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"23c7904ad4905889eb3b476b6012e61cd0db81357dc282484c1273f664551e4a"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204261 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"a59c2cd2a86e447386b98c8ae0b549d25e103c6e0aa54f07054070409d8496ee"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204275 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"a977125b67b9294bb56e9b61199b7602c12da5c9c2332514a161007f18a9303c"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204309 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"dbf34e412a3edcde991c53a218c3ca5c91ab2414ac507c8adb456ed9fa61e44e"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.204324 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"333fbdf93c4f3c0f5b37ac3c5311ed6b367753cf5d440e8f958c6e712bb2e4d1"} Mar 13 14:22:18 crc kubenswrapper[4907]: I0313 14:22:18.236135 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.219919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6r77v" event={"ID":"f5203ff6-d473-484d-ab8f-414e9ca804d4","Type":"ContainerStarted","Data":"02fa6003694dc80306fe15a5f7194247a3867e3bda8fe6e237d269d53553a6a0"} Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.220478 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.223528 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736"} Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.259739 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6r77v" podStartSLOduration=5.991184115 podStartE2EDuration="13.25971355s" podCreationTimestamp="2026-03-13 14:22:06 +0000 UTC" firstStartedPulling="2026-03-13 14:22:07.480587326 +0000 UTC m=+1026.380375015" lastFinishedPulling="2026-03-13 14:22:14.749116761 +0000 UTC m=+1033.648904450" observedRunningTime="2026-03-13 14:22:19.254451896 +0000 UTC m=+1038.154239595" watchObservedRunningTime="2026-03-13 14:22:19.25971355 +0000 UTC m=+1038.159501239" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.333793 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:19 crc kubenswrapper[4907]: I0313 14:22:19.380586 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:21 crc kubenswrapper[4907]: I0313 14:22:21.239123 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d4jn9" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" containerID="cri-o://a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a" gracePeriod=2 Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.245475 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.250010 4907 generic.go:334] "Generic (PLEG): container finished" podID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerID="a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a" exitCode=0 Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.250069 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a"} Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.288661 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.487464 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.614340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") pod \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.614855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") pod \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.614947 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") pod \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\" (UID: \"4211d5e0-93f1-4c48-a60e-6c89ae200a59\") " Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.615826 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities" (OuterVolumeSpecName: "utilities") pod "4211d5e0-93f1-4c48-a60e-6c89ae200a59" (UID: "4211d5e0-93f1-4c48-a60e-6c89ae200a59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.620581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs" (OuterVolumeSpecName: "kube-api-access-95pxs") pod "4211d5e0-93f1-4c48-a60e-6c89ae200a59" (UID: "4211d5e0-93f1-4c48-a60e-6c89ae200a59"). InnerVolumeSpecName "kube-api-access-95pxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.672862 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4211d5e0-93f1-4c48-a60e-6c89ae200a59" (UID: "4211d5e0-93f1-4c48-a60e-6c89ae200a59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.716373 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.716431 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95pxs\" (UniqueName: \"kubernetes.io/projected/4211d5e0-93f1-4c48-a60e-6c89ae200a59-kube-api-access-95pxs\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:22 crc kubenswrapper[4907]: I0313 14:22:22.716452 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4211d5e0-93f1-4c48-a60e-6c89ae200a59-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.264641 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4jn9" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.264637 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4jn9" event={"ID":"4211d5e0-93f1-4c48-a60e-6c89ae200a59","Type":"ContainerDied","Data":"9ed471a362cc5821dd762507d4b70cdc1ef3e27ed3258ebfcd19ca5d91f1ce3f"} Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.264810 4907 scope.go:117] "RemoveContainer" containerID="a91b547a2dfe97aa6285764503c56a2bb34d57df00449c14921d3d722377c73a" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.289439 4907 scope.go:117] "RemoveContainer" containerID="1006ba55273fa47807d756205c115736def78e9d78f5d0bc2e7b68771fa457bf" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.326346 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.333595 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d4jn9"] Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.346312 4907 scope.go:117] "RemoveContainer" containerID="4cc7ca19806ba6c30546ebc2e565b8d92a5693e0b133afb20dfc043acb245e70" Mar 13 14:22:23 crc kubenswrapper[4907]: I0313 14:22:23.799528 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" path="/var/lib/kubelet/pods/4211d5e0-93f1-4c48-a60e-6c89ae200a59/volumes" Mar 13 14:22:27 crc kubenswrapper[4907]: I0313 14:22:27.242455 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-jlsw9" Mar 13 14:22:27 crc kubenswrapper[4907]: I0313 14:22:27.247213 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6r77v" Mar 13 14:22:27 crc kubenswrapper[4907]: I0313 14:22:27.932761 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-2g4kt" Mar 13 14:22:28 crc kubenswrapper[4907]: I0313 14:22:28.818287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-fkplv" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302437 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g"] Mar 13 14:22:30 crc kubenswrapper[4907]: E0313 14:22:30.302652 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302664 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" Mar 13 14:22:30 crc kubenswrapper[4907]: E0313 14:22:30.302674 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-utilities" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302680 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-utilities" Mar 13 14:22:30 crc kubenswrapper[4907]: E0313 14:22:30.302689 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-content" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="extract-content" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.302791 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4211d5e0-93f1-4c48-a60e-6c89ae200a59" containerName="registry-server" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.303586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.305526 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.313630 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g"] Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.352438 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.352779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.352809 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.453758 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.453806 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.453838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.454353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.454346 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.473297 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.637580 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:30 crc kubenswrapper[4907]: I0313 14:22:30.873632 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g"] Mar 13 14:22:31 crc kubenswrapper[4907]: I0313 14:22:31.321529 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerID="9a7f8034ae10429297ca47eef0ac8cf7185b4b0681d362f7ac3ab16e8b94f648" exitCode=0 Mar 13 14:22:31 crc kubenswrapper[4907]: I0313 14:22:31.321566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"9a7f8034ae10429297ca47eef0ac8cf7185b4b0681d362f7ac3ab16e8b94f648"} Mar 13 14:22:31 crc kubenswrapper[4907]: I0313 14:22:31.321590 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerStarted","Data":"015cef428accbaa410a9449f28eee6690287054668a794e7e0447f58adb4d487"} Mar 13 14:22:34 crc kubenswrapper[4907]: I0313 14:22:34.339747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerStarted","Data":"79c7a18f632a438161553f4588a9282777d07b1799c22f74e34574606d71fc57"} Mar 13 14:22:35 crc kubenswrapper[4907]: I0313 14:22:35.348613 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerID="79c7a18f632a438161553f4588a9282777d07b1799c22f74e34574606d71fc57" exitCode=0 Mar 13 14:22:35 crc kubenswrapper[4907]: I0313 14:22:35.348700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"79c7a18f632a438161553f4588a9282777d07b1799c22f74e34574606d71fc57"} Mar 13 14:22:36 crc kubenswrapper[4907]: I0313 14:22:36.356962 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerID="5ad71ff77a738ebdca98e6804ade91d8b524277872fa6f4892e2b3d140b3443f" exitCode=0 Mar 13 14:22:36 crc kubenswrapper[4907]: I0313 14:22:36.357082 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"5ad71ff77a738ebdca98e6804ade91d8b524277872fa6f4892e2b3d140b3443f"} Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.629598 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.669995 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") pod \"1ba163f7-0dce-42a8-ab55-15aba1940065\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.670084 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") pod \"1ba163f7-0dce-42a8-ab55-15aba1940065\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.670185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") pod \"1ba163f7-0dce-42a8-ab55-15aba1940065\" (UID: \"1ba163f7-0dce-42a8-ab55-15aba1940065\") " Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.672929 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle" (OuterVolumeSpecName: "bundle") pod "1ba163f7-0dce-42a8-ab55-15aba1940065" (UID: "1ba163f7-0dce-42a8-ab55-15aba1940065"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.678766 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk" (OuterVolumeSpecName: "kube-api-access-xg5pk") pod "1ba163f7-0dce-42a8-ab55-15aba1940065" (UID: "1ba163f7-0dce-42a8-ab55-15aba1940065"). InnerVolumeSpecName "kube-api-access-xg5pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.684056 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util" (OuterVolumeSpecName: "util") pod "1ba163f7-0dce-42a8-ab55-15aba1940065" (UID: "1ba163f7-0dce-42a8-ab55-15aba1940065"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.772213 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.772275 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg5pk\" (UniqueName: \"kubernetes.io/projected/1ba163f7-0dce-42a8-ab55-15aba1940065-kube-api-access-xg5pk\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:37 crc kubenswrapper[4907]: I0313 14:22:37.772292 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1ba163f7-0dce-42a8-ab55-15aba1940065-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:22:38 crc kubenswrapper[4907]: I0313 14:22:38.369339 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" event={"ID":"1ba163f7-0dce-42a8-ab55-15aba1940065","Type":"ContainerDied","Data":"015cef428accbaa410a9449f28eee6690287054668a794e7e0447f58adb4d487"} Mar 13 14:22:38 crc kubenswrapper[4907]: I0313 14:22:38.369394 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="015cef428accbaa410a9449f28eee6690287054668a794e7e0447f58adb4d487" Mar 13 14:22:38 crc kubenswrapper[4907]: I0313 14:22:38.369414 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.422519 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj"] Mar 13 14:22:43 crc kubenswrapper[4907]: E0313 14:22:43.423518 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="util" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423531 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="util" Mar 13 14:22:43 crc kubenswrapper[4907]: E0313 14:22:43.423553 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="pull" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423560 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="pull" Mar 13 14:22:43 crc kubenswrapper[4907]: E0313 14:22:43.423571 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="extract" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423579 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="extract" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.423708 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ba163f7-0dce-42a8-ab55-15aba1940065" containerName="extract" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.424232 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.426753 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.426824 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.427107 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-mjvjd" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.445628 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj"] Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.567901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7ff33217-23de-4659-8503-ba87078b122c-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.567972 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vdsn\" (UniqueName: \"kubernetes.io/projected/7ff33217-23de-4659-8503-ba87078b122c-kube-api-access-5vdsn\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.669820 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7ff33217-23de-4659-8503-ba87078b122c-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.669899 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vdsn\" (UniqueName: \"kubernetes.io/projected/7ff33217-23de-4659-8503-ba87078b122c-kube-api-access-5vdsn\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.670642 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7ff33217-23de-4659-8503-ba87078b122c-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.694227 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vdsn\" (UniqueName: \"kubernetes.io/projected/7ff33217-23de-4659-8503-ba87078b122c-kube-api-access-5vdsn\") pod \"cert-manager-operator-controller-manager-66c8bdd694-jzdbj\" (UID: \"7ff33217-23de-4659-8503-ba87078b122c\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:43 crc kubenswrapper[4907]: I0313 14:22:43.747341 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" Mar 13 14:22:44 crc kubenswrapper[4907]: I0313 14:22:44.237955 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj"] Mar 13 14:22:44 crc kubenswrapper[4907]: W0313 14:22:44.246520 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ff33217_23de_4659_8503_ba87078b122c.slice/crio-89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701 WatchSource:0}: Error finding container 89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701: Status 404 returned error can't find the container with id 89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701 Mar 13 14:22:44 crc kubenswrapper[4907]: I0313 14:22:44.403838 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" event={"ID":"7ff33217-23de-4659-8503-ba87078b122c","Type":"ContainerStarted","Data":"89165b1c30c4ec07deb4d7feafe9b15d6512ed2dfa67ff11476f2079553d3701"} Mar 13 14:22:48 crc kubenswrapper[4907]: I0313 14:22:48.431225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" event={"ID":"7ff33217-23de-4659-8503-ba87078b122c","Type":"ContainerStarted","Data":"244d6d5079fbb4cedc0ad864a08b16f344431675a94efbb7435358a38cc6d135"} Mar 13 14:22:48 crc kubenswrapper[4907]: I0313 14:22:48.454052 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-jzdbj" podStartSLOduration=2.21627938 podStartE2EDuration="5.454020321s" podCreationTimestamp="2026-03-13 14:22:43 +0000 UTC" firstStartedPulling="2026-03-13 14:22:44.248243039 +0000 UTC m=+1063.148030728" lastFinishedPulling="2026-03-13 14:22:47.48598398 +0000 UTC m=+1066.385771669" observedRunningTime="2026-03-13 14:22:48.448745764 +0000 UTC m=+1067.348533453" watchObservedRunningTime="2026-03-13 14:22:48.454020321 +0000 UTC m=+1067.353808020" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.421817 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-tk9wv"] Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.423462 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.426658 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.429279 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.433189 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-tk9wv"] Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.433373 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-nm4p6" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.481223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.481284 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmqf\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-kube-api-access-kgmqf\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.582354 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.582431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmqf\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-kube-api-access-kgmqf\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.607326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.607328 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmqf\" (UniqueName: \"kubernetes.io/projected/3634e065-182f-49b7-af9c-01ea4779b1a6-kube-api-access-kgmqf\") pod \"cert-manager-webhook-6888856db4-tk9wv\" (UID: \"3634e065-182f-49b7-af9c-01ea4779b1a6\") " pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:51 crc kubenswrapper[4907]: I0313 14:22:51.745460 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:52 crc kubenswrapper[4907]: I0313 14:22:52.245934 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-tk9wv"] Mar 13 14:22:52 crc kubenswrapper[4907]: I0313 14:22:52.453914 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" event={"ID":"3634e065-182f-49b7-af9c-01ea4779b1a6","Type":"ContainerStarted","Data":"093ceea01d9d08e0fe67f760877688892a998bd62331ee65bcfaba6a5f019313"} Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.846987 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-v92wj"] Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.847654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.851217 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tk6cd" Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.859547 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-v92wj"] Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.927994 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:54 crc kubenswrapper[4907]: I0313 14:22:54.928113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlr7g\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-kube-api-access-jlr7g\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.030101 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.030207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlr7g\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-kube-api-access-jlr7g\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.054110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.054999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlr7g\" (UniqueName: \"kubernetes.io/projected/e1799412-2b44-413c-8c9d-b362432fd536-kube-api-access-jlr7g\") pod \"cert-manager-cainjector-5545bd876-v92wj\" (UID: \"e1799412-2b44-413c-8c9d-b362432fd536\") " pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.223209 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" Mar 13 14:22:55 crc kubenswrapper[4907]: I0313 14:22:55.631684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-v92wj"] Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.555484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" event={"ID":"e1799412-2b44-413c-8c9d-b362432fd536","Type":"ContainerStarted","Data":"6d73d860d2e2c77085d118696a4d100a2ba5c83d6d87b67fd063290b86d671f3"} Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.556663 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" event={"ID":"3634e065-182f-49b7-af9c-01ea4779b1a6","Type":"ContainerStarted","Data":"ce77e2ddb9406e42c13eea05c925689ed9ac92bd180f84bf11d1e3f23c548581"} Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.557035 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:22:57 crc kubenswrapper[4907]: I0313 14:22:57.590809 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" podStartSLOduration=1.86427175 podStartE2EDuration="6.590793842s" podCreationTimestamp="2026-03-13 14:22:51 +0000 UTC" firstStartedPulling="2026-03-13 14:22:52.26275769 +0000 UTC m=+1071.162545379" lastFinishedPulling="2026-03-13 14:22:56.989279782 +0000 UTC m=+1075.889067471" observedRunningTime="2026-03-13 14:22:57.586380569 +0000 UTC m=+1076.486168258" watchObservedRunningTime="2026-03-13 14:22:57.590793842 +0000 UTC m=+1076.490581531" Mar 13 14:22:58 crc kubenswrapper[4907]: I0313 14:22:58.562934 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" event={"ID":"e1799412-2b44-413c-8c9d-b362432fd536","Type":"ContainerStarted","Data":"f705f01b0f23fb4e01b414400e8950c0fdce958711b3dd3e51bab6e21366b850"} Mar 13 14:22:58 crc kubenswrapper[4907]: I0313 14:22:58.581212 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-v92wj" podStartSLOduration=3.517988513 podStartE2EDuration="4.581191407s" podCreationTimestamp="2026-03-13 14:22:54 +0000 UTC" firstStartedPulling="2026-03-13 14:22:56.926660227 +0000 UTC m=+1075.826447916" lastFinishedPulling="2026-03-13 14:22:57.989863091 +0000 UTC m=+1076.889650810" observedRunningTime="2026-03-13 14:22:58.575859428 +0000 UTC m=+1077.475647117" watchObservedRunningTime="2026-03-13 14:22:58.581191407 +0000 UTC m=+1077.480979096" Mar 13 14:23:02 crc kubenswrapper[4907]: I0313 14:23:02.592679 4907 scope.go:117] "RemoveContainer" containerID="ee5ef069f4132418de863818eaddffd619fd31484fffe36ce7827c7d59840bbd" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.771799 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-rn7sp"] Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.773429 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.778974 4907 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jvdvb" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.787409 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-rn7sp"] Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.888966 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-bound-sa-token\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.889241 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tx9wq\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-kube-api-access-tx9wq\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.990449 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tx9wq\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-kube-api-access-tx9wq\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:04 crc kubenswrapper[4907]: I0313 14:23:04.990535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-bound-sa-token\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.009840 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tx9wq\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-kube-api-access-tx9wq\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.010830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8a238ffb-ab19-472f-bb14-580e63f0fb7d-bound-sa-token\") pod \"cert-manager-545d4d4674-rn7sp\" (UID: \"8a238ffb-ab19-472f-bb14-580e63f0fb7d\") " pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.122797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-rn7sp" Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.549188 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-rn7sp"] Mar 13 14:23:05 crc kubenswrapper[4907]: W0313 14:23:05.561271 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a238ffb_ab19_472f_bb14_580e63f0fb7d.slice/crio-584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d WatchSource:0}: Error finding container 584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d: Status 404 returned error can't find the container with id 584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d Mar 13 14:23:05 crc kubenswrapper[4907]: I0313 14:23:05.605231 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-rn7sp" event={"ID":"8a238ffb-ab19-472f-bb14-580e63f0fb7d","Type":"ContainerStarted","Data":"584f553807d7e58f6c1ea9e6145adedf6646747aea75f7a1e516745d22377d5d"} Mar 13 14:23:06 crc kubenswrapper[4907]: I0313 14:23:06.612853 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-rn7sp" event={"ID":"8a238ffb-ab19-472f-bb14-580e63f0fb7d","Type":"ContainerStarted","Data":"dd446bcce87574a0ca09e0574b1f618887d39ad0c4423f66b4b23d04fccda7c1"} Mar 13 14:23:06 crc kubenswrapper[4907]: I0313 14:23:06.626823 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-rn7sp" podStartSLOduration=2.626804645 podStartE2EDuration="2.626804645s" podCreationTimestamp="2026-03-13 14:23:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:23:06.62590826 +0000 UTC m=+1085.525695979" watchObservedRunningTime="2026-03-13 14:23:06.626804645 +0000 UTC m=+1085.526592334" Mar 13 14:23:06 crc kubenswrapper[4907]: I0313 14:23:06.772472 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-tk9wv" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.971152 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.972598 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.975369 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-5ljjn" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.975924 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 13 14:23:09 crc kubenswrapper[4907]: I0313 14:23:09.978116 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.028385 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.065640 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"openstack-operator-index-sp5p9\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.166757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"openstack-operator-index-sp5p9\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.188608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"openstack-operator-index-sp5p9\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.295375 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.515090 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:10 crc kubenswrapper[4907]: I0313 14:23:10.656256 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerStarted","Data":"0f9b7376b543a8fb6b597bc980448ce89cc3f22d712c728980f748145ec9f64b"} Mar 13 14:23:11 crc kubenswrapper[4907]: I0313 14:23:11.669577 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerStarted","Data":"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528"} Mar 13 14:23:11 crc kubenswrapper[4907]: I0313 14:23:11.687224 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-sp5p9" podStartSLOduration=1.9511680120000001 podStartE2EDuration="2.68720042s" podCreationTimestamp="2026-03-13 14:23:09 +0000 UTC" firstStartedPulling="2026-03-13 14:23:10.522080156 +0000 UTC m=+1089.421867845" lastFinishedPulling="2026-03-13 14:23:11.258112564 +0000 UTC m=+1090.157900253" observedRunningTime="2026-03-13 14:23:11.681974894 +0000 UTC m=+1090.581762593" watchObservedRunningTime="2026-03-13 14:23:11.68720042 +0000 UTC m=+1090.586988149" Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.350818 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.679323 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-sp5p9" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" containerID="cri-o://35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" gracePeriod=2 Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.961069 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-z98nv"] Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.962683 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:13 crc kubenswrapper[4907]: I0313 14:23:13.985640 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z98nv"] Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.016369 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prsvc\" (UniqueName: \"kubernetes.io/projected/c23245c8-0fba-483e-b126-09349d413030-kube-api-access-prsvc\") pod \"openstack-operator-index-z98nv\" (UID: \"c23245c8-0fba-483e-b126-09349d413030\") " pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.086121 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.117298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prsvc\" (UniqueName: \"kubernetes.io/projected/c23245c8-0fba-483e-b126-09349d413030-kube-api-access-prsvc\") pod \"openstack-operator-index-z98nv\" (UID: \"c23245c8-0fba-483e-b126-09349d413030\") " pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.139827 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prsvc\" (UniqueName: \"kubernetes.io/projected/c23245c8-0fba-483e-b126-09349d413030-kube-api-access-prsvc\") pod \"openstack-operator-index-z98nv\" (UID: \"c23245c8-0fba-483e-b126-09349d413030\") " pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.218808 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") pod \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\" (UID: \"919e211f-e4a5-4af6-8ae8-81e2500ab50b\") " Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.223417 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp" (OuterVolumeSpecName: "kube-api-access-bnxsp") pod "919e211f-e4a5-4af6-8ae8-81e2500ab50b" (UID: "919e211f-e4a5-4af6-8ae8-81e2500ab50b"). InnerVolumeSpecName "kube-api-access-bnxsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.305353 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.320176 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnxsp\" (UniqueName: \"kubernetes.io/projected/919e211f-e4a5-4af6-8ae8-81e2500ab50b-kube-api-access-bnxsp\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688083 4907 generic.go:334] "Generic (PLEG): container finished" podID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" exitCode=0 Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688561 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sp5p9" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerDied","Data":"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528"} Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688706 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sp5p9" event={"ID":"919e211f-e4a5-4af6-8ae8-81e2500ab50b","Type":"ContainerDied","Data":"0f9b7376b543a8fb6b597bc980448ce89cc3f22d712c728980f748145ec9f64b"} Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.688756 4907 scope.go:117] "RemoveContainer" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.717837 4907 scope.go:117] "RemoveContainer" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" Mar 13 14:23:14 crc kubenswrapper[4907]: E0313 14:23:14.718748 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528\": container with ID starting with 35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528 not found: ID does not exist" containerID="35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.718821 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528"} err="failed to get container status \"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528\": rpc error: code = NotFound desc = could not find container \"35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528\": container with ID starting with 35fa0a9a1b21ee3f2bcb0895ae1cf3a5cf49c10c24083a3e95693b10237b2528 not found: ID does not exist" Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.722482 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.728793 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-sp5p9"] Mar 13 14:23:14 crc kubenswrapper[4907]: I0313 14:23:14.763223 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z98nv"] Mar 13 14:23:14 crc kubenswrapper[4907]: W0313 14:23:14.769732 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc23245c8_0fba_483e_b126_09349d413030.slice/crio-e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86 WatchSource:0}: Error finding container e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86: Status 404 returned error can't find the container with id e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86 Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.700174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z98nv" event={"ID":"c23245c8-0fba-483e-b126-09349d413030","Type":"ContainerStarted","Data":"5b6ec09d34114f19900bf9c8b582572254b25c91acb3e2a5a15a991df195c79b"} Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.700836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z98nv" event={"ID":"c23245c8-0fba-483e-b126-09349d413030","Type":"ContainerStarted","Data":"e06b2c4cd29ed453c67abc329e6e12a8b3bb3370b0dd96514d62c69eca51be86"} Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.725625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-z98nv" podStartSLOduration=2.236400147 podStartE2EDuration="2.725584888s" podCreationTimestamp="2026-03-13 14:23:13 +0000 UTC" firstStartedPulling="2026-03-13 14:23:14.774060526 +0000 UTC m=+1093.673848215" lastFinishedPulling="2026-03-13 14:23:15.263245277 +0000 UTC m=+1094.163032956" observedRunningTime="2026-03-13 14:23:15.712782092 +0000 UTC m=+1094.612569781" watchObservedRunningTime="2026-03-13 14:23:15.725584888 +0000 UTC m=+1094.625372617" Mar 13 14:23:15 crc kubenswrapper[4907]: I0313 14:23:15.793604 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" path="/var/lib/kubelet/pods/919e211f-e4a5-4af6-8ae8-81e2500ab50b/volumes" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.366704 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:20 crc kubenswrapper[4907]: E0313 14:23:20.368179 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.368207 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.368427 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="919e211f-e4a5-4af6-8ae8-81e2500ab50b" containerName="registry-server" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.370489 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.376369 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.432480 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.433065 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.433148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.534776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.534919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.535066 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.535690 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.535824 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.567918 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"redhat-marketplace-4tdsc\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:20 crc kubenswrapper[4907]: I0313 14:23:20.702832 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.196838 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.765196 4907 generic.go:334] "Generic (PLEG): container finished" podID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" exitCode=0 Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.765283 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711"} Mar 13 14:23:21 crc kubenswrapper[4907]: I0313 14:23:21.765338 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerStarted","Data":"3ee4542adf3dfb6a230d733a6eecb4aea1254a119b039f5276071b36fe25d5af"} Mar 13 14:23:23 crc kubenswrapper[4907]: I0313 14:23:23.785488 4907 generic.go:334] "Generic (PLEG): container finished" podID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" exitCode=0 Mar 13 14:23:23 crc kubenswrapper[4907]: I0313 14:23:23.797733 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b"} Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.305567 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.305686 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.353288 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.797002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerStarted","Data":"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d"} Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.848415 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-z98nv" Mar 13 14:23:24 crc kubenswrapper[4907]: I0313 14:23:24.852497 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4tdsc" podStartSLOduration=2.285607132 podStartE2EDuration="4.852451159s" podCreationTimestamp="2026-03-13 14:23:20 +0000 UTC" firstStartedPulling="2026-03-13 14:23:21.767909439 +0000 UTC m=+1100.667697128" lastFinishedPulling="2026-03-13 14:23:24.334753456 +0000 UTC m=+1103.234541155" observedRunningTime="2026-03-13 14:23:24.827251607 +0000 UTC m=+1103.727039336" watchObservedRunningTime="2026-03-13 14:23:24.852451159 +0000 UTC m=+1103.752238868" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.602798 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv"] Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.603942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.608874 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-kzqgh" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.617754 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv"] Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.723524 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.723641 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.723694 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.825693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.826362 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.826627 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.827085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.827726 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.855726 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:25 crc kubenswrapper[4907]: I0313 14:23:25.952762 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.434168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv"] Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.812787 4907 generic.go:334] "Generic (PLEG): container finished" podID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerID="56614d64f72203351cf4d876cc8dfb5b9282687bb09eeafa3abb1fdbd52371d1" exitCode=0 Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.812848 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"56614d64f72203351cf4d876cc8dfb5b9282687bb09eeafa3abb1fdbd52371d1"} Mar 13 14:23:26 crc kubenswrapper[4907]: I0313 14:23:26.813582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerStarted","Data":"f7a55efc6d71185f02a71437939fed4071aac8483a065cea96bb2638c19b8a80"} Mar 13 14:23:29 crc kubenswrapper[4907]: I0313 14:23:29.832627 4907 generic.go:334] "Generic (PLEG): container finished" podID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerID="99a13d1aea20b5d81fd740254033bb43e952a711596b576c4767e172d46118a5" exitCode=0 Mar 13 14:23:29 crc kubenswrapper[4907]: I0313 14:23:29.832687 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"99a13d1aea20b5d81fd740254033bb43e952a711596b576c4767e172d46118a5"} Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.703733 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.704096 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.760430 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.842965 4907 generic.go:334] "Generic (PLEG): container finished" podID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerID="ca026a81b43b3f27ef8a94825e51af847ca34ff14e2cf7bf77947bc3fc682487" exitCode=0 Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.843033 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"ca026a81b43b3f27ef8a94825e51af847ca34ff14e2cf7bf77947bc3fc682487"} Mar 13 14:23:30 crc kubenswrapper[4907]: I0313 14:23:30.891162 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.115313 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.117475 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") pod \"29e80c51-f1f5-4850-8e81-ee52fe060a43\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.117621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") pod \"29e80c51-f1f5-4850-8e81-ee52fe060a43\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.117713 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") pod \"29e80c51-f1f5-4850-8e81-ee52fe060a43\" (UID: \"29e80c51-f1f5-4850-8e81-ee52fe060a43\") " Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.118850 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle" (OuterVolumeSpecName: "bundle") pod "29e80c51-f1f5-4850-8e81-ee52fe060a43" (UID: "29e80c51-f1f5-4850-8e81-ee52fe060a43"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.125157 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv" (OuterVolumeSpecName: "kube-api-access-8wlxv") pod "29e80c51-f1f5-4850-8e81-ee52fe060a43" (UID: "29e80c51-f1f5-4850-8e81-ee52fe060a43"). InnerVolumeSpecName "kube-api-access-8wlxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.219063 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wlxv\" (UniqueName: \"kubernetes.io/projected/29e80c51-f1f5-4850-8e81-ee52fe060a43-kube-api-access-8wlxv\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.219103 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.299233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util" (OuterVolumeSpecName: "util") pod "29e80c51-f1f5-4850-8e81-ee52fe060a43" (UID: "29e80c51-f1f5-4850-8e81-ee52fe060a43"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.321570 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/29e80c51-f1f5-4850-8e81-ee52fe060a43-util\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.353063 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.865766 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" event={"ID":"29e80c51-f1f5-4850-8e81-ee52fe060a43","Type":"ContainerDied","Data":"f7a55efc6d71185f02a71437939fed4071aac8483a065cea96bb2638c19b8a80"} Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.865843 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7a55efc6d71185f02a71437939fed4071aac8483a065cea96bb2638c19b8a80" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.865867 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv" Mar 13 14:23:32 crc kubenswrapper[4907]: I0313 14:23:32.866077 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4tdsc" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" containerID="cri-o://a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" gracePeriod=2 Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.251695 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.334951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") pod \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.335025 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") pod \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.335097 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") pod \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\" (UID: \"0a48608e-b0d0-4ff5-963f-0fc5fae809d5\") " Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.336131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities" (OuterVolumeSpecName: "utilities") pod "0a48608e-b0d0-4ff5-963f-0fc5fae809d5" (UID: "0a48608e-b0d0-4ff5-963f-0fc5fae809d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.345288 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c" (OuterVolumeSpecName: "kube-api-access-m474c") pod "0a48608e-b0d0-4ff5-963f-0fc5fae809d5" (UID: "0a48608e-b0d0-4ff5-963f-0fc5fae809d5"). InnerVolumeSpecName "kube-api-access-m474c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.372187 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a48608e-b0d0-4ff5-963f-0fc5fae809d5" (UID: "0a48608e-b0d0-4ff5-963f-0fc5fae809d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.436238 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m474c\" (UniqueName: \"kubernetes.io/projected/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-kube-api-access-m474c\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.436282 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.436295 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a48608e-b0d0-4ff5-963f-0fc5fae809d5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873294 4907 generic.go:334] "Generic (PLEG): container finished" podID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" exitCode=0 Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873505 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d"} Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873579 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4tdsc" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873591 4907 scope.go:117] "RemoveContainer" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.873576 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4tdsc" event={"ID":"0a48608e-b0d0-4ff5-963f-0fc5fae809d5","Type":"ContainerDied","Data":"3ee4542adf3dfb6a230d733a6eecb4aea1254a119b039f5276071b36fe25d5af"} Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.891412 4907 scope.go:117] "RemoveContainer" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.892152 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.896871 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4tdsc"] Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.912234 4907 scope.go:117] "RemoveContainer" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.926330 4907 scope.go:117] "RemoveContainer" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" Mar 13 14:23:33 crc kubenswrapper[4907]: E0313 14:23:33.927430 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d\": container with ID starting with a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d not found: ID does not exist" containerID="a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927460 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d"} err="failed to get container status \"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d\": rpc error: code = NotFound desc = could not find container \"a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d\": container with ID starting with a2bf587b73502d20df6faba1a74fb50eb91bfbf9b9fa5e9342562ea96082f40d not found: ID does not exist" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927483 4907 scope.go:117] "RemoveContainer" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" Mar 13 14:23:33 crc kubenswrapper[4907]: E0313 14:23:33.927811 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b\": container with ID starting with 859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b not found: ID does not exist" containerID="859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927834 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b"} err="failed to get container status \"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b\": rpc error: code = NotFound desc = could not find container \"859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b\": container with ID starting with 859e57ce5122694ec07507ac13d55d3387cef8b8951ec996ebb84737ad268a3b not found: ID does not exist" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.927851 4907 scope.go:117] "RemoveContainer" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" Mar 13 14:23:33 crc kubenswrapper[4907]: E0313 14:23:33.929339 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711\": container with ID starting with e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711 not found: ID does not exist" containerID="e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711" Mar 13 14:23:33 crc kubenswrapper[4907]: I0313 14:23:33.929363 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711"} err="failed to get container status \"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711\": rpc error: code = NotFound desc = could not find container \"e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711\": container with ID starting with e917c4d1488ff6e34a1eae85a4179e14a1fb7da267d8c17bc59f188d9f93a711 not found: ID does not exist" Mar 13 14:23:35 crc kubenswrapper[4907]: I0313 14:23:35.791762 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" path="/var/lib/kubelet/pods/0a48608e-b0d0-4ff5-963f-0fc5fae809d5/volumes" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.647713 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr"] Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648420 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="util" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648468 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="util" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648497 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-content" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648511 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-content" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648537 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="pull" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648552 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="pull" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648574 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="extract" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648589 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="extract" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648618 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-utilities" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648638 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="extract-utilities" Mar 13 14:23:36 crc kubenswrapper[4907]: E0313 14:23:36.648698 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648712 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.648992 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="29e80c51-f1f5-4850-8e81-ee52fe060a43" containerName="extract" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.649040 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a48608e-b0d0-4ff5-963f-0fc5fae809d5" containerName="registry-server" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.649668 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.652005 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-slwjf" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.667189 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr"] Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.678502 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpcx5\" (UniqueName: \"kubernetes.io/projected/dc51eaa2-313d-4256-a145-3b73b02e2dfe-kube-api-access-kpcx5\") pod \"openstack-operator-controller-init-6dc56d8cd6-hxpjr\" (UID: \"dc51eaa2-313d-4256-a145-3b73b02e2dfe\") " pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.779858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpcx5\" (UniqueName: \"kubernetes.io/projected/dc51eaa2-313d-4256-a145-3b73b02e2dfe-kube-api-access-kpcx5\") pod \"openstack-operator-controller-init-6dc56d8cd6-hxpjr\" (UID: \"dc51eaa2-313d-4256-a145-3b73b02e2dfe\") " pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.803057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpcx5\" (UniqueName: \"kubernetes.io/projected/dc51eaa2-313d-4256-a145-3b73b02e2dfe-kube-api-access-kpcx5\") pod \"openstack-operator-controller-init-6dc56d8cd6-hxpjr\" (UID: \"dc51eaa2-313d-4256-a145-3b73b02e2dfe\") " pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:36 crc kubenswrapper[4907]: I0313 14:23:36.973817 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:37 crc kubenswrapper[4907]: I0313 14:23:37.424341 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr"] Mar 13 14:23:37 crc kubenswrapper[4907]: I0313 14:23:37.904404 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" event={"ID":"dc51eaa2-313d-4256-a145-3b73b02e2dfe","Type":"ContainerStarted","Data":"024e1e4036febf3a5b23ce7ce85464428178f1fb3ec1537da29c5d516178dad4"} Mar 13 14:23:41 crc kubenswrapper[4907]: I0313 14:23:41.935434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" event={"ID":"dc51eaa2-313d-4256-a145-3b73b02e2dfe","Type":"ContainerStarted","Data":"c6bd1731c8cc406b0d04b52d5d560717b4f7b481fb734a5174b56fb30c9b9f82"} Mar 13 14:23:41 crc kubenswrapper[4907]: I0313 14:23:41.936228 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:23:41 crc kubenswrapper[4907]: I0313 14:23:41.972692 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" podStartSLOduration=2.303056543 podStartE2EDuration="5.972675457s" podCreationTimestamp="2026-03-13 14:23:36 +0000 UTC" firstStartedPulling="2026-03-13 14:23:37.433175397 +0000 UTC m=+1116.332963086" lastFinishedPulling="2026-03-13 14:23:41.102794311 +0000 UTC m=+1120.002582000" observedRunningTime="2026-03-13 14:23:41.967548574 +0000 UTC m=+1120.867336263" watchObservedRunningTime="2026-03-13 14:23:41.972675457 +0000 UTC m=+1120.872463146" Mar 13 14:23:46 crc kubenswrapper[4907]: I0313 14:23:46.975990 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-6dc56d8cd6-hxpjr" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.136794 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.138089 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.140411 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.140635 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.141059 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.142357 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.194347 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"auto-csr-approver-29556864-768cg\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.295197 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"auto-csr-approver-29556864-768cg\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.311038 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"auto-csr-approver-29556864-768cg\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.454507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:00 crc kubenswrapper[4907]: I0313 14:24:00.937232 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:24:01 crc kubenswrapper[4907]: I0313 14:24:01.051545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556864-768cg" event={"ID":"32f92b7e-159a-4824-93e7-ea6d49961db8","Type":"ContainerStarted","Data":"6cc6b642475e740e71d15bc948ee2b3420da216f5037a2a22c34dcc911c783d7"} Mar 13 14:24:03 crc kubenswrapper[4907]: I0313 14:24:03.073688 4907 generic.go:334] "Generic (PLEG): container finished" podID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerID="a53552e35726d3cc22de750fc04d8a15ada7be07f868ed527291e30242034c7d" exitCode=0 Mar 13 14:24:03 crc kubenswrapper[4907]: I0313 14:24:03.073764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556864-768cg" event={"ID":"32f92b7e-159a-4824-93e7-ea6d49961db8","Type":"ContainerDied","Data":"a53552e35726d3cc22de750fc04d8a15ada7be07f868ed527291e30242034c7d"} Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.366332 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.471849 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") pod \"32f92b7e-159a-4824-93e7-ea6d49961db8\" (UID: \"32f92b7e-159a-4824-93e7-ea6d49961db8\") " Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.477496 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz" (OuterVolumeSpecName: "kube-api-access-2nhqz") pod "32f92b7e-159a-4824-93e7-ea6d49961db8" (UID: "32f92b7e-159a-4824-93e7-ea6d49961db8"). InnerVolumeSpecName "kube-api-access-2nhqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:24:04 crc kubenswrapper[4907]: I0313 14:24:04.573155 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nhqz\" (UniqueName: \"kubernetes.io/projected/32f92b7e-159a-4824-93e7-ea6d49961db8-kube-api-access-2nhqz\") on node \"crc\" DevicePath \"\"" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.087502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556864-768cg" event={"ID":"32f92b7e-159a-4824-93e7-ea6d49961db8","Type":"ContainerDied","Data":"6cc6b642475e740e71d15bc948ee2b3420da216f5037a2a22c34dcc911c783d7"} Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.087802 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cc6b642475e740e71d15bc948ee2b3420da216f5037a2a22c34dcc911c783d7" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.087580 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556864-768cg" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.408744 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.412917 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556858-bk6rj"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.793858 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a" path="/var/lib/kubelet/pods/03fa7dad-f13d-4e5c-8b9c-60b2ef2d272a/volumes" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.884686 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4"] Mar 13 14:24:05 crc kubenswrapper[4907]: E0313 14:24:05.885028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerName="oc" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.885052 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerName="oc" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.885214 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" containerName="oc" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.885766 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.887680 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xftd5" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.899073 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.899833 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.904323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-f2bvz" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.909864 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.915559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.924149 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.924978 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.927794 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hfwlj" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.932480 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gll76"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.933296 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.937322 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-84xsv" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.953349 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gll76"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.959963 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.987851 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4zll\" (UniqueName: \"kubernetes.io/projected/863c77e9-4022-4ab9-8ac7-a4006d22813c-kube-api-access-t4zll\") pod \"barbican-operator-controller-manager-d47688694-tl8t4\" (UID: \"863c77e9-4022-4ab9-8ac7-a4006d22813c\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988062 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gv2c\" (UniqueName: \"kubernetes.io/projected/b7890be5-5773-4562-b1ab-8e647aa16256-kube-api-access-5gv2c\") pod \"glance-operator-controller-manager-5964f64c48-gll76\" (UID: \"b7890be5-5773-4562-b1ab-8e647aa16256\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988133 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltpg6\" (UniqueName: \"kubernetes.io/projected/77b1eb71-a455-4ec3-8e17-a847575cf0b0-kube-api-access-ltpg6\") pod \"cinder-operator-controller-manager-984cd4dcf-w497d\" (UID: \"77b1eb71-a455-4ec3-8e17-a847575cf0b0\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988209 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk68h\" (UniqueName: \"kubernetes.io/projected/8789d213-46a4-4649-add6-51ebe4412c27-kube-api-access-tk68h\") pod \"designate-operator-controller-manager-66d56f6ff4-7cbb2\" (UID: \"8789d213-46a4-4649-add6-51ebe4412c27\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.988799 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-874m2"] Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.989540 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:05 crc kubenswrapper[4907]: I0313 14:24:05.992243 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-jfmkp" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.016252 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-874m2"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.023140 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.023894 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.027318 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-6tth2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.029183 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.029940 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.033792 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-5tgz9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.034091 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.045335 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.046269 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.048489 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-swjxd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.048653 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.060677 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.076196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.087051 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.087934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.092258 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6tq54" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gv2c\" (UniqueName: \"kubernetes.io/projected/b7890be5-5773-4562-b1ab-8e647aa16256-kube-api-access-5gv2c\") pod \"glance-operator-controller-manager-5964f64c48-gll76\" (UID: \"b7890be5-5773-4562-b1ab-8e647aa16256\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093316 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltpg6\" (UniqueName: \"kubernetes.io/projected/77b1eb71-a455-4ec3-8e17-a847575cf0b0-kube-api-access-ltpg6\") pod \"cinder-operator-controller-manager-984cd4dcf-w497d\" (UID: \"77b1eb71-a455-4ec3-8e17-a847575cf0b0\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093337 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4whsv\" (UniqueName: \"kubernetes.io/projected/64d55221-60c5-4521-ac84-beed4e9b3993-kube-api-access-4whsv\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmv9x\" (UniqueName: \"kubernetes.io/projected/8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66-kube-api-access-dmv9x\") pod \"horizon-operator-controller-manager-6d9d6b584d-r7bhq\" (UID: \"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093388 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk68h\" (UniqueName: \"kubernetes.io/projected/8789d213-46a4-4649-add6-51ebe4412c27-kube-api-access-tk68h\") pod \"designate-operator-controller-manager-66d56f6ff4-7cbb2\" (UID: \"8789d213-46a4-4649-add6-51ebe4412c27\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64424\" (UniqueName: \"kubernetes.io/projected/d3cdd2ca-3e81-480f-876e-2b975d8bc9d8-kube-api-access-64424\") pod \"heat-operator-controller-manager-77b6666d85-874m2\" (UID: \"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093450 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4zll\" (UniqueName: \"kubernetes.io/projected/863c77e9-4022-4ab9-8ac7-a4006d22813c-kube-api-access-t4zll\") pod \"barbican-operator-controller-manager-d47688694-tl8t4\" (UID: \"863c77e9-4022-4ab9-8ac7-a4006d22813c\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.093478 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59tqw\" (UniqueName: \"kubernetes.io/projected/19c320fa-6619-45cd-a73e-65834173b786-kube-api-access-59tqw\") pod \"ironic-operator-controller-manager-5bc894d9b-w5829\" (UID: \"19c320fa-6619-45cd-a73e-65834173b786\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.162513 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gv2c\" (UniqueName: \"kubernetes.io/projected/b7890be5-5773-4562-b1ab-8e647aa16256-kube-api-access-5gv2c\") pod \"glance-operator-controller-manager-5964f64c48-gll76\" (UID: \"b7890be5-5773-4562-b1ab-8e647aa16256\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.175043 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.185968 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk68h\" (UniqueName: \"kubernetes.io/projected/8789d213-46a4-4649-add6-51ebe4412c27-kube-api-access-tk68h\") pod \"designate-operator-controller-manager-66d56f6ff4-7cbb2\" (UID: \"8789d213-46a4-4649-add6-51ebe4412c27\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.192679 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltpg6\" (UniqueName: \"kubernetes.io/projected/77b1eb71-a455-4ec3-8e17-a847575cf0b0-kube-api-access-ltpg6\") pod \"cinder-operator-controller-manager-984cd4dcf-w497d\" (UID: \"77b1eb71-a455-4ec3-8e17-a847575cf0b0\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.208258 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4zll\" (UniqueName: \"kubernetes.io/projected/863c77e9-4022-4ab9-8ac7-a4006d22813c-kube-api-access-t4zll\") pod \"barbican-operator-controller-manager-d47688694-tl8t4\" (UID: \"863c77e9-4022-4ab9-8ac7-a4006d22813c\") " pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212741 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64424\" (UniqueName: \"kubernetes.io/projected/d3cdd2ca-3e81-480f-876e-2b975d8bc9d8-kube-api-access-64424\") pod \"heat-operator-controller-manager-77b6666d85-874m2\" (UID: \"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212805 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7mn9\" (UniqueName: \"kubernetes.io/projected/7d620a90-0e93-41dc-ab39-fc6ea8c461e7-kube-api-access-n7mn9\") pod \"keystone-operator-controller-manager-684f77d66d-7rdzd\" (UID: \"7d620a90-0e93-41dc-ab39-fc6ea8c461e7\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59tqw\" (UniqueName: \"kubernetes.io/projected/19c320fa-6619-45cd-a73e-65834173b786-kube-api-access-59tqw\") pod \"ironic-operator-controller-manager-5bc894d9b-w5829\" (UID: \"19c320fa-6619-45cd-a73e-65834173b786\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212914 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4whsv\" (UniqueName: \"kubernetes.io/projected/64d55221-60c5-4521-ac84-beed4e9b3993-kube-api-access-4whsv\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.212932 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmv9x\" (UniqueName: \"kubernetes.io/projected/8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66-kube-api-access-dmv9x\") pod \"horizon-operator-controller-manager-6d9d6b584d-r7bhq\" (UID: \"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.213553 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.213603 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:06.713587441 +0000 UTC m=+1145.613375130 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.214017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.229148 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.263723 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.264500 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.264930 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64424\" (UniqueName: \"kubernetes.io/projected/d3cdd2ca-3e81-480f-876e-2b975d8bc9d8-kube-api-access-64424\") pod \"heat-operator-controller-manager-77b6666d85-874m2\" (UID: \"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.265012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4whsv\" (UniqueName: \"kubernetes.io/projected/64d55221-60c5-4521-ac84-beed4e9b3993-kube-api-access-4whsv\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.265279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmv9x\" (UniqueName: \"kubernetes.io/projected/8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66-kube-api-access-dmv9x\") pod \"horizon-operator-controller-manager-6d9d6b584d-r7bhq\" (UID: \"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.265586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.268074 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.268793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59tqw\" (UniqueName: \"kubernetes.io/projected/19c320fa-6619-45cd-a73e-65834173b786-kube-api-access-59tqw\") pod \"ironic-operator-controller-manager-5bc894d9b-w5829\" (UID: \"19c320fa-6619-45cd-a73e-65834173b786\") " pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.269110 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.272934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.278160 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-x5qcj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.280767 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-hj4pg" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.283243 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.285710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.287126 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fwlbf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.304085 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.311554 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314180 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b25c5\" (UniqueName: \"kubernetes.io/projected/29201cb7-c056-4e4e-b0da-cc775c7ef11f-kube-api-access-b25c5\") pod \"neutron-operator-controller-manager-776c5696bf-gvgdl\" (UID: \"29201cb7-c056-4e4e-b0da-cc775c7ef11f\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7mn9\" (UniqueName: \"kubernetes.io/projected/7d620a90-0e93-41dc-ab39-fc6ea8c461e7-kube-api-access-n7mn9\") pod \"keystone-operator-controller-manager-684f77d66d-7rdzd\" (UID: \"7d620a90-0e93-41dc-ab39-fc6ea8c461e7\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314288 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmthx\" (UniqueName: \"kubernetes.io/projected/e8134e91-4b9d-4e3c-9aec-eba18acabdce-kube-api-access-nmthx\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw\" (UID: \"e8134e91-4b9d-4e3c-9aec-eba18acabdce\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.314314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cq7h\" (UniqueName: \"kubernetes.io/projected/c55515f1-509e-4a4e-813d-fbc466fca489-kube-api-access-7cq7h\") pod \"manila-operator-controller-manager-57b484b4df-m8j7m\" (UID: \"c55515f1-509e-4a4e-813d-fbc466fca489\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.317251 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.323733 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.337447 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.337844 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7mn9\" (UniqueName: \"kubernetes.io/projected/7d620a90-0e93-41dc-ab39-fc6ea8c461e7-kube-api-access-n7mn9\") pod \"keystone-operator-controller-manager-684f77d66d-7rdzd\" (UID: \"7d620a90-0e93-41dc-ab39-fc6ea8c461e7\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.338255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.340061 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-67nq5" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.343660 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.344510 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.345796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.346590 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-kcfs6" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.352332 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.376272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.385930 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.394092 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.394990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.399200 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.399583 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-q8r6n" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.415513 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416579 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b25c5\" (UniqueName: \"kubernetes.io/projected/29201cb7-c056-4e4e-b0da-cc775c7ef11f-kube-api-access-b25c5\") pod \"neutron-operator-controller-manager-776c5696bf-gvgdl\" (UID: \"29201cb7-c056-4e4e-b0da-cc775c7ef11f\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gkcl\" (UniqueName: \"kubernetes.io/projected/4ab8f569-de56-4fea-9be0-07908063522a-kube-api-access-2gkcl\") pod \"nova-operator-controller-manager-7f84474648-ql4kd\" (UID: \"4ab8f569-de56-4fea-9be0-07908063522a\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416641 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xtmd\" (UniqueName: \"kubernetes.io/projected/f915558a-c0ab-4d8e-a427-e56a7588e382-kube-api-access-5xtmd\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416677 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416710 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmthx\" (UniqueName: \"kubernetes.io/projected/e8134e91-4b9d-4e3c-9aec-eba18acabdce-kube-api-access-nmthx\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw\" (UID: \"e8134e91-4b9d-4e3c-9aec-eba18acabdce\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416735 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cq7h\" (UniqueName: \"kubernetes.io/projected/c55515f1-509e-4a4e-813d-fbc466fca489-kube-api-access-7cq7h\") pod \"manila-operator-controller-manager-57b484b4df-m8j7m\" (UID: \"c55515f1-509e-4a4e-813d-fbc466fca489\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.416801 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lwf7\" (UniqueName: \"kubernetes.io/projected/5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473-kube-api-access-2lwf7\") pod \"octavia-operator-controller-manager-5f4f55cb5c-sk2l9\" (UID: \"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.428573 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.449773 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b25c5\" (UniqueName: \"kubernetes.io/projected/29201cb7-c056-4e4e-b0da-cc775c7ef11f-kube-api-access-b25c5\") pod \"neutron-operator-controller-manager-776c5696bf-gvgdl\" (UID: \"29201cb7-c056-4e4e-b0da-cc775c7ef11f\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.449828 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmthx\" (UniqueName: \"kubernetes.io/projected/e8134e91-4b9d-4e3c-9aec-eba18acabdce-kube-api-access-nmthx\") pod \"mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw\" (UID: \"e8134e91-4b9d-4e3c-9aec-eba18acabdce\") " pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.456265 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.457101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.460440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-znvp6" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.462824 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cq7h\" (UniqueName: \"kubernetes.io/projected/c55515f1-509e-4a4e-813d-fbc466fca489-kube-api-access-7cq7h\") pod \"manila-operator-controller-manager-57b484b4df-m8j7m\" (UID: \"c55515f1-509e-4a4e-813d-fbc466fca489\") " pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.485406 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.486403 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.488712 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jwrkw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.496757 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520033 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwf7\" (UniqueName: \"kubernetes.io/projected/5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473-kube-api-access-2lwf7\") pod \"octavia-operator-controller-manager-5f4f55cb5c-sk2l9\" (UID: \"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gkcl\" (UniqueName: \"kubernetes.io/projected/4ab8f569-de56-4fea-9be0-07908063522a-kube-api-access-2gkcl\") pod \"nova-operator-controller-manager-7f84474648-ql4kd\" (UID: \"4ab8f569-de56-4fea-9be0-07908063522a\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520318 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xtmd\" (UniqueName: \"kubernetes.io/projected/f915558a-c0ab-4d8e-a427-e56a7588e382-kube-api-access-5xtmd\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.520353 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.520523 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.520574 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.020556034 +0000 UTC m=+1145.920343733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.536945 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.547438 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xtmd\" (UniqueName: \"kubernetes.io/projected/f915558a-c0ab-4d8e-a427-e56a7588e382-kube-api-access-5xtmd\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.552939 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.560928 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.554485 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwf7\" (UniqueName: \"kubernetes.io/projected/5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473-kube-api-access-2lwf7\") pod \"octavia-operator-controller-manager-5f4f55cb5c-sk2l9\" (UID: \"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.562675 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gkcl\" (UniqueName: \"kubernetes.io/projected/4ab8f569-de56-4fea-9be0-07908063522a-kube-api-access-2gkcl\") pod \"nova-operator-controller-manager-7f84474648-ql4kd\" (UID: \"4ab8f569-de56-4fea-9be0-07908063522a\") " pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.572060 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-6kktj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.586944 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.592542 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.610421 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.611182 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.622426 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzlzx\" (UniqueName: \"kubernetes.io/projected/fdddbc99-1a71-4250-ab2a-6f426e744423-kube-api-access-mzlzx\") pod \"ovn-operator-controller-manager-bbc5b68f9-cqmhw\" (UID: \"fdddbc99-1a71-4250-ab2a-6f426e744423\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.622528 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5trp\" (UniqueName: \"kubernetes.io/projected/0b3db736-b3d4-494f-9755-9103150d267f-kube-api-access-d5trp\") pod \"placement-operator-controller-manager-574d45c66c-b99f7\" (UID: \"0b3db736-b3d4-494f-9755-9103150d267f\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.629873 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.631778 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-pcq8d" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.636259 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.665738 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.665901 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.679274 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.689933 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.690781 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.691751 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.710588 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bjcb6" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.717818 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.718798 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.725099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sp8zt" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730591 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwz7r\" (UniqueName: \"kubernetes.io/projected/d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1-kube-api-access-mwz7r\") pod \"telemetry-operator-controller-manager-6854b8b9d9-dj95f\" (UID: \"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpzf7\" (UniqueName: \"kubernetes.io/projected/92718823-f344-4d7b-beff-e75f1574f1e2-kube-api-access-hpzf7\") pod \"swift-operator-controller-manager-7f9cc5dd44-sd6nr\" (UID: \"92718823-f344-4d7b-beff-e75f1574f1e2\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5trp\" (UniqueName: \"kubernetes.io/projected/0b3db736-b3d4-494f-9755-9103150d267f-kube-api-access-d5trp\") pod \"placement-operator-controller-manager-574d45c66c-b99f7\" (UID: \"0b3db736-b3d4-494f-9755-9103150d267f\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.730747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzlzx\" (UniqueName: \"kubernetes.io/projected/fdddbc99-1a71-4250-ab2a-6f426e744423-kube-api-access-mzlzx\") pod \"ovn-operator-controller-manager-bbc5b68f9-cqmhw\" (UID: \"fdddbc99-1a71-4250-ab2a-6f426e744423\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.731355 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: E0313 14:24:06.731778 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.731756759 +0000 UTC m=+1146.631544448 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.743386 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.791646 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5trp\" (UniqueName: \"kubernetes.io/projected/0b3db736-b3d4-494f-9755-9103150d267f-kube-api-access-d5trp\") pod \"placement-operator-controller-manager-574d45c66c-b99f7\" (UID: \"0b3db736-b3d4-494f-9755-9103150d267f\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.793420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzlzx\" (UniqueName: \"kubernetes.io/projected/fdddbc99-1a71-4250-ab2a-6f426e744423-kube-api-access-mzlzx\") pod \"ovn-operator-controller-manager-bbc5b68f9-cqmhw\" (UID: \"fdddbc99-1a71-4250-ab2a-6f426e744423\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.834782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpzf7\" (UniqueName: \"kubernetes.io/projected/92718823-f344-4d7b-beff-e75f1574f1e2-kube-api-access-hpzf7\") pod \"swift-operator-controller-manager-7f9cc5dd44-sd6nr\" (UID: \"92718823-f344-4d7b-beff-e75f1574f1e2\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.834861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7lx\" (UniqueName: \"kubernetes.io/projected/a89fbf9b-b391-426b-9f9d-91cdba71602b-kube-api-access-5h7lx\") pod \"watcher-operator-controller-manager-6c4d75f7f9-hjp4p\" (UID: \"a89fbf9b-b391-426b-9f9d-91cdba71602b\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.834925 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4b47\" (UniqueName: \"kubernetes.io/projected/a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe-kube-api-access-b4b47\") pod \"test-operator-controller-manager-5c5cb9c4d7-vzpcd\" (UID: \"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.835020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwz7r\" (UniqueName: \"kubernetes.io/projected/d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1-kube-api-access-mwz7r\") pod \"telemetry-operator-controller-manager-6854b8b9d9-dj95f\" (UID: \"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.836318 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.866607 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.878211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpzf7\" (UniqueName: \"kubernetes.io/projected/92718823-f344-4d7b-beff-e75f1574f1e2-kube-api-access-hpzf7\") pod \"swift-operator-controller-manager-7f9cc5dd44-sd6nr\" (UID: \"92718823-f344-4d7b-beff-e75f1574f1e2\") " pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.879064 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwz7r\" (UniqueName: \"kubernetes.io/projected/d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1-kube-api-access-mwz7r\") pod \"telemetry-operator-controller-manager-6854b8b9d9-dj95f\" (UID: \"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1\") " pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.904613 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.905780 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910085 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910290 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910438 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-xm9nq" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.910791 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.922388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.927098 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.938328 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.939232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7lx\" (UniqueName: \"kubernetes.io/projected/a89fbf9b-b391-426b-9f9d-91cdba71602b-kube-api-access-5h7lx\") pod \"watcher-operator-controller-manager-6c4d75f7f9-hjp4p\" (UID: \"a89fbf9b-b391-426b-9f9d-91cdba71602b\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.947196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4b47\" (UniqueName: \"kubernetes.io/projected/a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe-kube-api-access-b4b47\") pod \"test-operator-controller-manager-5c5cb9c4d7-vzpcd\" (UID: \"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.950314 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb"] Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.950682 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.964767 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-nn9ml" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.966563 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.967330 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4b47\" (UniqueName: \"kubernetes.io/projected/a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe-kube-api-access-b4b47\") pod \"test-operator-controller-manager-5c5cb9c4d7-vzpcd\" (UID: \"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:06 crc kubenswrapper[4907]: I0313 14:24:06.991420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7lx\" (UniqueName: \"kubernetes.io/projected/a89fbf9b-b391-426b-9f9d-91cdba71602b-kube-api-access-5h7lx\") pod \"watcher-operator-controller-manager-6c4d75f7f9-hjp4p\" (UID: \"a89fbf9b-b391-426b-9f9d-91cdba71602b\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp585\" (UniqueName: \"kubernetes.io/projected/44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d-kube-api-access-kp585\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtvzb\" (UID: \"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047684 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047757 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047780 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.047804 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfzck\" (UniqueName: \"kubernetes.io/projected/9ac47edf-03ab-4e93-b238-2aa00765ab06-kube-api-access-bfzck\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.047961 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.048014 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:08.04799598 +0000 UTC m=+1146.947783669 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.086912 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.095848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.109605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" event={"ID":"863c77e9-4022-4ab9-8ac7-a4006d22813c","Type":"ContainerStarted","Data":"bef85b99ba0d7433d5ec268d18ee5ad44e31169909fdb11b1722a7603775c913"} Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.134669 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148736 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148812 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfzck\" (UniqueName: \"kubernetes.io/projected/9ac47edf-03ab-4e93-b238-2aa00765ab06-kube-api-access-bfzck\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.148841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp585\" (UniqueName: \"kubernetes.io/projected/44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d-kube-api-access-kp585\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtvzb\" (UID: \"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.150605 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.150678 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.65066066 +0000 UTC m=+1146.550448349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.151534 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.151567 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:07.651557886 +0000 UTC m=+1146.551345575 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.181365 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfzck\" (UniqueName: \"kubernetes.io/projected/9ac47edf-03ab-4e93-b238-2aa00765ab06-kube-api-access-bfzck\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.186525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp585\" (UniqueName: \"kubernetes.io/projected/44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d-kube-api-access-kp585\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xtvzb\" (UID: \"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.298961 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.376390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-gll76"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.390062 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.656953 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.661717 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.664599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.664645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664892 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664916 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664954 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:08.66493421 +0000 UTC m=+1147.564721959 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.664975 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:08.664965941 +0000 UTC m=+1147.564753740 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.672805 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829"] Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.689591 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ab8f569_de56_4fea_9be0_07908063522a.slice/crio-2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88 WatchSource:0}: Error finding container 2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88: Status 404 returned error can't find the container with id 2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88 Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.689661 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.697443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.702846 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-874m2"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.708430 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq"] Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.726930 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19c320fa_6619_45cd_a73e_65834173b786.slice/crio-d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8 WatchSource:0}: Error finding container d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8: Status 404 returned error can't find the container with id d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8 Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.727857 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ecaf4b8_808e_44b9_8ae7_6ef9519c4a66.slice/crio-d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e WatchSource:0}: Error finding container d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e: Status 404 returned error can't find the container with id d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.767834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.768989 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.769053 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:09.76903617 +0000 UTC m=+1148.668823859 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.822495 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.829546 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.847036 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7"] Mar 13 14:24:07 crc kubenswrapper[4907]: I0313 14:24:07.851610 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl"] Mar 13 14:24:07 crc kubenswrapper[4907]: W0313 14:24:07.858133 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc55515f1_509e_4a4e_813d_fbc466fca489.slice/crio-f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59 WatchSource:0}: Error finding container f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59: Status 404 returned error can't find the container with id f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59 Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.868360 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d5trp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-b99f7_openstack-operators(0b3db736-b3d4-494f-9755-9103150d267f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:07 crc kubenswrapper[4907]: E0313 14:24:07.869797 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podUID="0b3db736-b3d4-494f-9755-9103150d267f" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.049381 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd"] Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.071983 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr"] Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.085287 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p"] Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.086897 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.087187 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.087258 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:10.087234075 +0000 UTC m=+1148.987021774 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.101371 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:72db77c98e7bca64d469b4dc316e9c8d329681f825d19ef8f333437fb1c6d3f5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hpzf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-7f9cc5dd44-sd6nr_openstack-operators(92718823-f344-4d7b-beff-e75f1574f1e2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.102722 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podUID="92718823-f344-4d7b-beff-e75f1574f1e2" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.103653 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5h7lx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-hjp4p_openstack-operators(a89fbf9b-b391-426b-9f9d-91cdba71602b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: W0313 14:24:08.103993 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdddbc99_1a71_4250_ab2a_6f426e744423.slice/crio-23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503 WatchSource:0}: Error finding container 23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503: Status 404 returned error can't find the container with id 23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503 Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.105206 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podUID="a89fbf9b-b391-426b-9f9d-91cdba71602b" Mar 13 14:24:08 crc kubenswrapper[4907]: W0313 14:24:08.105845 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44ecb24b_4c5c_4cf6_96e9_4d8b52469d3d.slice/crio-a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d WatchSource:0}: Error finding container a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d: Status 404 returned error can't find the container with id a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.106337 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mzlzx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bbc5b68f9-cqmhw_openstack-operators(fdddbc99-1a71-4250-ab2a-6f426e744423): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.107474 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podUID="fdddbc99-1a71-4250-ab2a-6f426e744423" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.107635 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kp585,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-xtvzb_openstack-operators(44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.108273 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw"] Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.110311 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podUID="44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.132493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" event={"ID":"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d","Type":"ContainerStarted","Data":"a8124cbf931bb327fe1039e05d89d2da5fa96c7a2fbfa266e53a0d33a1439c5d"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.135035 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb"] Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.135205 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podUID="44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.136043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" event={"ID":"77b1eb71-a455-4ec3-8e17-a847575cf0b0","Type":"ContainerStarted","Data":"8133a4150e9ff6586b24c5809c839ec3151e63009c4e9d4f463d2426457e2024"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.140267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" event={"ID":"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66","Type":"ContainerStarted","Data":"d023d3167efff6988171f93e3761488c1ce0729bd3cb6e8aa0a553e1a18fe33e"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.142790 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" event={"ID":"0b3db736-b3d4-494f-9755-9103150d267f","Type":"ContainerStarted","Data":"31ac27c8d988f95e7b59f94abe8de0c0755f0ea00708ce916ac4e0f6c75f2dd0"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.144118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podUID="0b3db736-b3d4-494f-9755-9103150d267f" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.169649 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" event={"ID":"19c320fa-6619-45cd-a73e-65834173b786","Type":"ContainerStarted","Data":"d18a8174c620b5a6ee3028caa18c2a577a0e97ec87adc0db628b0d1c7e9036a8"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.173546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" event={"ID":"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1","Type":"ContainerStarted","Data":"249124a101607411664b00fc3ece797bc08d43cce738dd15d70fdceaa982f6d3"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.188089 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" event={"ID":"c55515f1-509e-4a4e-813d-fbc466fca489","Type":"ContainerStarted","Data":"f3fff2cc89f13cc8f5f42b57d76dd61682015600446bfbf3151c19fd65c90e59"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.194143 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" event={"ID":"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe","Type":"ContainerStarted","Data":"079300236dd122d951f49a408fe129275af832b44c6f7e64416f2a0469ccba77"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.195847 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" event={"ID":"4ab8f569-de56-4fea-9be0-07908063522a","Type":"ContainerStarted","Data":"2bed19672f429f39772b19fc86747e4dbbce4b48f9ab5aac93c9a3800c55fd88"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.197298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" event={"ID":"b7890be5-5773-4562-b1ab-8e647aa16256","Type":"ContainerStarted","Data":"24a301c9f2dbd19e528ed3398be77fecd32947246f3a9d49c7babaca70800a42"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.198480 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" event={"ID":"7d620a90-0e93-41dc-ab39-fc6ea8c461e7","Type":"ContainerStarted","Data":"439611099e1dbdaaa0b92adbf75b260031ff4a85f58ee41465805b2d5cfbd289"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.205805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" event={"ID":"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8","Type":"ContainerStarted","Data":"7effad99e4ddab39739a3036988a870d9b248b765b7764f466306ec247ad5974"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.208043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" event={"ID":"a89fbf9b-b391-426b-9f9d-91cdba71602b","Type":"ContainerStarted","Data":"6734e50dcc6089686b369fd6c9ffe2bbbb6b6683f81309eeb5d5494d6365b007"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.209137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" event={"ID":"29201cb7-c056-4e4e-b0da-cc775c7ef11f","Type":"ContainerStarted","Data":"730ba7fd8e2c2f7a3080b0b3b70054d68651814baacb28b06daf6c30ac9c2710"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.210529 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podUID="a89fbf9b-b391-426b-9f9d-91cdba71602b" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.216399 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" event={"ID":"8789d213-46a4-4649-add6-51ebe4412c27","Type":"ContainerStarted","Data":"542f7823b45dccbadb548ebf3a6b53de33234194035260a112d7af8fa2a6c417"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.218679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" event={"ID":"92718823-f344-4d7b-beff-e75f1574f1e2","Type":"ContainerStarted","Data":"471369223f2f7aeca31c0e6f7d51eb344c95995296092c5fe268a7081154592c"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.220142 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" event={"ID":"e8134e91-4b9d-4e3c-9aec-eba18acabdce","Type":"ContainerStarted","Data":"5b5bc5bef33ef19b061bdb6b2f2e5536bd5cb5c80c42fa20343f63c0e712ee5b"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.220255 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72db77c98e7bca64d469b4dc316e9c8d329681f825d19ef8f333437fb1c6d3f5\\\"\"" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podUID="92718823-f344-4d7b-beff-e75f1574f1e2" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.221559 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" event={"ID":"fdddbc99-1a71-4250-ab2a-6f426e744423","Type":"ContainerStarted","Data":"23437d6b081e76de7b1c3ba9d56430532e367089e5a7f5e6435afb9e118f4503"} Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.222673 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podUID="fdddbc99-1a71-4250-ab2a-6f426e744423" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.223198 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" event={"ID":"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473","Type":"ContainerStarted","Data":"d870537e08a01962b5adaee2ef2e6fa2ced689caa2a1f2199d1293e680e16ad7"} Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.701483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:08 crc kubenswrapper[4907]: I0313 14:24:08.701803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.702036 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.702112 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:10.702094757 +0000 UTC m=+1149.601882436 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.703930 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:08 crc kubenswrapper[4907]: E0313 14:24:08.703988 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:10.70397205 +0000 UTC m=+1149.603759739 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240350 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podUID="a89fbf9b-b391-426b-9f9d-91cdba71602b" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240675 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podUID="0b3db736-b3d4-494f-9755-9103150d267f" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240713 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podUID="44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.240740 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:72db77c98e7bca64d469b4dc316e9c8d329681f825d19ef8f333437fb1c6d3f5\\\"\"" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podUID="92718823-f344-4d7b-beff-e75f1574f1e2" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.250084 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podUID="fdddbc99-1a71-4250-ab2a-6f426e744423" Mar 13 14:24:09 crc kubenswrapper[4907]: I0313 14:24:09.826637 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.826859 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:09 crc kubenswrapper[4907]: E0313 14:24:09.826926 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:13.826912147 +0000 UTC m=+1152.726699836 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: I0313 14:24:10.135847 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.136586 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.136648 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:14.136625616 +0000 UTC m=+1153.036413315 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: I0313 14:24:10.747575 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:10 crc kubenswrapper[4907]: I0313 14:24:10.747753 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748007 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748063 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:14.748048022 +0000 UTC m=+1153.647835711 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748104 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:10 crc kubenswrapper[4907]: E0313 14:24:10.748123 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:14.748117114 +0000 UTC m=+1153.647904803 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:13 crc kubenswrapper[4907]: I0313 14:24:13.894822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:13 crc kubenswrapper[4907]: E0313 14:24:13.895220 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:13 crc kubenswrapper[4907]: E0313 14:24:13.895266 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:21.89525322 +0000 UTC m=+1160.795040909 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: I0313 14:24:14.236451 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.236635 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.236738 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:22.236708674 +0000 UTC m=+1161.136496363 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: I0313 14:24:14.844722 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:14 crc kubenswrapper[4907]: I0313 14:24:14.844786 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.844927 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.844994 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:22.844976502 +0000 UTC m=+1161.744764191 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.845032 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:14 crc kubenswrapper[4907]: E0313 14:24:14.845110 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:22.845092415 +0000 UTC m=+1161.744880104 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:18 crc kubenswrapper[4907]: I0313 14:24:18.041473 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:24:18 crc kubenswrapper[4907]: I0313 14:24:18.041977 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:24:20 crc kubenswrapper[4907]: E0313 14:24:20.901052 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721" Mar 13 14:24:20 crc kubenswrapper[4907]: E0313 14:24:20.901708 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b25c5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-776c5696bf-gvgdl_openstack-operators(29201cb7-c056-4e4e-b0da-cc775c7ef11f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:20 crc kubenswrapper[4907]: E0313 14:24:20.903132 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" podUID="29201cb7-c056-4e4e-b0da-cc775c7ef11f" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.357320 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" podUID="29201cb7-c056-4e4e-b0da-cc775c7ef11f" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.666391 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:a26d062af19b3bc6dc6633171f1eff8eec33e8e925465d4968a0b9a36012a7e7" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.666569 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:a26d062af19b3bc6dc6633171f1eff8eec33e8e925465d4968a0b9a36012a7e7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nmthx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw_openstack-operators(e8134e91-4b9d-4e3c-9aec-eba18acabdce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.667924 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" podUID="e8134e91-4b9d-4e3c-9aec-eba18acabdce" Mar 13 14:24:21 crc kubenswrapper[4907]: I0313 14:24:21.964996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.965652 4907 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:21 crc kubenswrapper[4907]: E0313 14:24:21.965701 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert podName:64d55221-60c5-4521-ac84-beed4e9b3993 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:37.96568585 +0000 UTC m=+1176.865473539 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert") pod "infra-operator-controller-manager-54dc5b8f8d-lp7kj" (UID: "64d55221-60c5-4521-ac84-beed4e9b3993") : secret "infra-operator-webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: I0313 14:24:22.270014 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.270296 4907 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.270358 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert podName:f915558a-c0ab-4d8e-a427-e56a7588e382 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:38.270343409 +0000 UTC m=+1177.170131098 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert") pod "openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" (UID: "f915558a-c0ab-4d8e-a427-e56a7588e382") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.340428 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:dd62e104225ea255af5a32828af4c21e1dfb50fbdf35cd41d07d1326f9017a40" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.340587 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:dd62e104225ea255af5a32828af4c21e1dfb50fbdf35cd41d07d1326f9017a40,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7cq7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-57b484b4df-m8j7m_openstack-operators(c55515f1-509e-4a4e-813d-fbc466fca489): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.342493 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" podUID="c55515f1-509e-4a4e-813d-fbc466fca489" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.366418 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:dd62e104225ea255af5a32828af4c21e1dfb50fbdf35cd41d07d1326f9017a40\\\"\"" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" podUID="c55515f1-509e-4a4e-813d-fbc466fca489" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.366633 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a26d062af19b3bc6dc6633171f1eff8eec33e8e925465d4968a0b9a36012a7e7\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" podUID="e8134e91-4b9d-4e3c-9aec-eba18acabdce" Mar 13 14:24:22 crc kubenswrapper[4907]: I0313 14:24:22.879281 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:22 crc kubenswrapper[4907]: I0313 14:24:22.879640 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.879500 4907 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.879973 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:38.879949903 +0000 UTC m=+1177.779737592 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "webhook-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.879993 4907 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 14:24:22 crc kubenswrapper[4907]: E0313 14:24:22.880037 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs podName:9ac47edf-03ab-4e93-b238-2aa00765ab06 nodeName:}" failed. No retries permitted until 2026-03-13 14:24:38.880025606 +0000 UTC m=+1177.779813395 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs") pod "openstack-operator-controller-manager-6484b7b757-qldxp" (UID: "9ac47edf-03ab-4e93-b238-2aa00765ab06") : secret "metrics-server-cert" not found Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.356564 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:bbe772fa051f782c9dcc3c34ce43495e1116aa9089a760c10068790baa9b25ff" Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.356739 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:bbe772fa051f782c9dcc3c34ce43495e1116aa9089a760c10068790baa9b25ff,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gkcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-7f84474648-ql4kd_openstack-operators(4ab8f569-de56-4fea-9be0-07908063522a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.357932 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" podUID="4ab8f569-de56-4fea-9be0-07908063522a" Mar 13 14:24:24 crc kubenswrapper[4907]: E0313 14:24:24.378895 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:bbe772fa051f782c9dcc3c34ce43495e1116aa9089a760c10068790baa9b25ff\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" podUID="4ab8f569-de56-4fea-9be0-07908063522a" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.409912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" event={"ID":"5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473","Type":"ContainerStarted","Data":"05abb31bbf4ec7fe38a62ff30c059ce6b087be181a0e28042622a070932727fd"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.410595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.411998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" event={"ID":"7d620a90-0e93-41dc-ab39-fc6ea8c461e7","Type":"ContainerStarted","Data":"93faad90b79e0bc4524f13f90e4ca5cc1287f3e4911bbcede638bdc9c21a3ee8"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.412580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.414507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" event={"ID":"a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe","Type":"ContainerStarted","Data":"512d1ab3188bf1032e497da8cbbffdc3e5d0fb10c97420ee700178ce6c2df88a"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.414936 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.416013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" event={"ID":"b7890be5-5773-4562-b1ab-8e647aa16256","Type":"ContainerStarted","Data":"43d4c09378c5dcd8bef4c18832d79b482b75c17adb040b1d95fb1f2993ee3ab2"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.416332 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.417391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" event={"ID":"92718823-f344-4d7b-beff-e75f1574f1e2","Type":"ContainerStarted","Data":"1426ea8e86e38c9303fa0dc9e6313ceb1f59abf0760282c44b6ec01bd4aae436"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.417720 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.419299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" event={"ID":"0b3db736-b3d4-494f-9755-9103150d267f","Type":"ContainerStarted","Data":"7d0eab1124c4f5223e3ff4691e59dd0f8e154658825f67875ee41d5b34768f3e"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.419605 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.421673 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" event={"ID":"8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66","Type":"ContainerStarted","Data":"75ae9afda13bc4d203b15ff875e1f989dc0aee5879dcf40a5388c3239c039ecb"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.421821 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.423025 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" event={"ID":"fdddbc99-1a71-4250-ab2a-6f426e744423","Type":"ContainerStarted","Data":"225ace33198a3da486e62fa067504b35a227f923a438459073f4ea5dc42a43c6"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.423175 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.424383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" event={"ID":"a89fbf9b-b391-426b-9f9d-91cdba71602b","Type":"ContainerStarted","Data":"7102de9beb3b8815117d2ed9366d7536e6f3f834b813e68b06eb8d8f1bffbf25"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.424595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.426165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" event={"ID":"44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d","Type":"ContainerStarted","Data":"0ae4f540d42cbc607b9ece7623709fa969fa343a1eabbbf6fe6be4e3fd1c241c"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.427563 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" event={"ID":"77b1eb71-a455-4ec3-8e17-a847575cf0b0","Type":"ContainerStarted","Data":"d5eb9803af60e9700eca1bfdb4a88a53f1da388443d89a4bcb492cf22e743924"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.427706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.429526 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" event={"ID":"d3cdd2ca-3e81-480f-876e-2b975d8bc9d8","Type":"ContainerStarted","Data":"e60ac8a501dd0be585475c0e81ac1273c6a26ae808f805e897cc13551694b004"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.429661 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.431420 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" event={"ID":"863c77e9-4022-4ab9-8ac7-a4006d22813c","Type":"ContainerStarted","Data":"35717b532a5717fb681c62aebea52a6a1b7aca58bed964725f877360feefee9f"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.431558 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.433095 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" event={"ID":"19c320fa-6619-45cd-a73e-65834173b786","Type":"ContainerStarted","Data":"090c2fbf2baac7dfb5c922f84512bb5616df8cf7281e1d6a388d38f2f4a6a2f9"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.433226 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.434628 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" event={"ID":"d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1","Type":"ContainerStarted","Data":"55330794c339c40fc7cd21675097faf304a29511453d047470018a57a2abdc0f"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.434745 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.436362 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" event={"ID":"8789d213-46a4-4649-add6-51ebe4412c27","Type":"ContainerStarted","Data":"898c1a1ae3b709eac2d6916ccecd30e0d7e2e968f67dc58e937fdda34ccc2478"} Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.437217 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.505506 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" podStartSLOduration=4.895257022 podStartE2EDuration="23.505477645s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.696858468 +0000 UTC m=+1146.596646167" lastFinishedPulling="2026-03-13 14:24:26.307079101 +0000 UTC m=+1165.206866790" observedRunningTime="2026-03-13 14:24:29.450874274 +0000 UTC m=+1168.350661963" watchObservedRunningTime="2026-03-13 14:24:29.505477645 +0000 UTC m=+1168.405265324" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.585641 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" podStartSLOduration=7.113427147 podStartE2EDuration="23.585611679s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.846129748 +0000 UTC m=+1146.745917437" lastFinishedPulling="2026-03-13 14:24:24.31831428 +0000 UTC m=+1163.218101969" observedRunningTime="2026-03-13 14:24:29.511649198 +0000 UTC m=+1168.411436887" watchObservedRunningTime="2026-03-13 14:24:29.585611679 +0000 UTC m=+1168.485399368" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.747944 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" podStartSLOduration=7.231655757 podStartE2EDuration="24.74792568s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.405503211 +0000 UTC m=+1146.305290900" lastFinishedPulling="2026-03-13 14:24:24.921773134 +0000 UTC m=+1163.821560823" observedRunningTime="2026-03-13 14:24:29.595635148 +0000 UTC m=+1168.495422837" watchObservedRunningTime="2026-03-13 14:24:29.74792568 +0000 UTC m=+1168.647713369" Mar 13 14:24:29 crc kubenswrapper[4907]: I0313 14:24:29.748762 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" podStartSLOduration=3.23411504 podStartE2EDuration="23.748757394s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.101212625 +0000 UTC m=+1147.001000314" lastFinishedPulling="2026-03-13 14:24:28.615854979 +0000 UTC m=+1167.515642668" observedRunningTime="2026-03-13 14:24:29.68255461 +0000 UTC m=+1168.582342299" watchObservedRunningTime="2026-03-13 14:24:29.748757394 +0000 UTC m=+1168.648545083" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.129325 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" podStartSLOduration=7.266576323 podStartE2EDuration="24.129308727s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.05760486 +0000 UTC m=+1146.957392549" lastFinishedPulling="2026-03-13 14:24:24.920337264 +0000 UTC m=+1163.820124953" observedRunningTime="2026-03-13 14:24:29.834851613 +0000 UTC m=+1168.734639302" watchObservedRunningTime="2026-03-13 14:24:30.129308727 +0000 UTC m=+1169.029096416" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.250709 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" podStartSLOduration=3.419635161 podStartE2EDuration="24.25068595s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.868166552 +0000 UTC m=+1146.767954241" lastFinishedPulling="2026-03-13 14:24:28.699217321 +0000 UTC m=+1167.599005030" observedRunningTime="2026-03-13 14:24:30.24427116 +0000 UTC m=+1169.144058849" watchObservedRunningTime="2026-03-13 14:24:30.25068595 +0000 UTC m=+1169.150473639" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.251410 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" podStartSLOduration=7.009688866 podStartE2EDuration="24.251403779s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.679619398 +0000 UTC m=+1146.579407087" lastFinishedPulling="2026-03-13 14:24:24.921334311 +0000 UTC m=+1163.821122000" observedRunningTime="2026-03-13 14:24:30.141099865 +0000 UTC m=+1169.040887554" watchObservedRunningTime="2026-03-13 14:24:30.251403779 +0000 UTC m=+1169.151191468" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.373509 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" podStartSLOduration=4.297184929 podStartE2EDuration="24.37349029s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.10352641 +0000 UTC m=+1147.003314089" lastFinishedPulling="2026-03-13 14:24:28.179831761 +0000 UTC m=+1167.079619450" observedRunningTime="2026-03-13 14:24:30.359562242 +0000 UTC m=+1169.259349931" watchObservedRunningTime="2026-03-13 14:24:30.37349029 +0000 UTC m=+1169.273277979" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.527920 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" podStartSLOduration=8.012141324 podStartE2EDuration="25.527902163s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.403790733 +0000 UTC m=+1146.303578422" lastFinishedPulling="2026-03-13 14:24:24.919551572 +0000 UTC m=+1163.819339261" observedRunningTime="2026-03-13 14:24:30.458097038 +0000 UTC m=+1169.357884727" watchObservedRunningTime="2026-03-13 14:24:30.527902163 +0000 UTC m=+1169.427689852" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.569601 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" podStartSLOduration=7.634635386 podStartE2EDuration="25.569583994s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:06.98626655 +0000 UTC m=+1145.886054229" lastFinishedPulling="2026-03-13 14:24:24.921215148 +0000 UTC m=+1163.821002837" observedRunningTime="2026-03-13 14:24:30.567315591 +0000 UTC m=+1169.467103291" watchObservedRunningTime="2026-03-13 14:24:30.569583994 +0000 UTC m=+1169.469371683" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.569713 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" podStartSLOduration=4.044845458 podStartE2EDuration="24.569709907s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.106246495 +0000 UTC m=+1147.006034184" lastFinishedPulling="2026-03-13 14:24:28.631110944 +0000 UTC m=+1167.530898633" observedRunningTime="2026-03-13 14:24:30.534740993 +0000 UTC m=+1169.434528682" watchObservedRunningTime="2026-03-13 14:24:30.569709907 +0000 UTC m=+1169.469497596" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.622028 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" podStartSLOduration=8.443666728 podStartE2EDuration="25.622006865s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.742992604 +0000 UTC m=+1146.642780293" lastFinishedPulling="2026-03-13 14:24:24.921332741 +0000 UTC m=+1163.821120430" observedRunningTime="2026-03-13 14:24:30.618129047 +0000 UTC m=+1169.517916746" watchObservedRunningTime="2026-03-13 14:24:30.622006865 +0000 UTC m=+1169.521794554" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.669216 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" podStartSLOduration=7.839257027 podStartE2EDuration="25.669198429s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.742903842 +0000 UTC m=+1146.642691531" lastFinishedPulling="2026-03-13 14:24:25.572845244 +0000 UTC m=+1164.472632933" observedRunningTime="2026-03-13 14:24:30.664145179 +0000 UTC m=+1169.563932868" watchObservedRunningTime="2026-03-13 14:24:30.669198429 +0000 UTC m=+1169.568986118" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.691498 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xtvzb" podStartSLOduration=4.061683358 podStartE2EDuration="24.69148052s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:08.107552832 +0000 UTC m=+1147.007340521" lastFinishedPulling="2026-03-13 14:24:28.737349994 +0000 UTC m=+1167.637137683" observedRunningTime="2026-03-13 14:24:30.689493725 +0000 UTC m=+1169.589281424" watchObservedRunningTime="2026-03-13 14:24:30.69148052 +0000 UTC m=+1169.591268209" Mar 13 14:24:30 crc kubenswrapper[4907]: I0313 14:24:30.722136 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" podStartSLOduration=7.984207326 podStartE2EDuration="25.722111694s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.182699673 +0000 UTC m=+1146.082487362" lastFinishedPulling="2026-03-13 14:24:24.920604041 +0000 UTC m=+1163.820391730" observedRunningTime="2026-03-13 14:24:30.719894793 +0000 UTC m=+1169.619682482" watchObservedRunningTime="2026-03-13 14:24:30.722111694 +0000 UTC m=+1169.621899373" Mar 13 14:24:32 crc kubenswrapper[4907]: I0313 14:24:32.805284 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" podStartSLOduration=11.229193399 podStartE2EDuration="27.805270126s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.744627169 +0000 UTC m=+1146.644414858" lastFinishedPulling="2026-03-13 14:24:24.320703896 +0000 UTC m=+1163.220491585" observedRunningTime="2026-03-13 14:24:30.756706978 +0000 UTC m=+1169.656494677" watchObservedRunningTime="2026-03-13 14:24:32.805270126 +0000 UTC m=+1171.705057815" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.219133 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-d47688694-tl8t4" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.239754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-w497d" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.266581 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-7cbb2" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.279595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-gll76" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.324657 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-874m2" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.352516 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-r7bhq" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.396386 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bc894d9b-w5829" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.421719 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-7rdzd" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.669108 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-sk2l9" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.839367 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-cqmhw" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.869250 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-b99f7" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.915969 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7f9cc5dd44-sd6nr" Mar 13 14:24:36 crc kubenswrapper[4907]: I0313 14:24:36.969743 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6854b8b9d9-dj95f" Mar 13 14:24:37 crc kubenswrapper[4907]: I0313 14:24:37.088899 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-vzpcd" Mar 13 14:24:37 crc kubenswrapper[4907]: I0313 14:24:37.098763 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-hjp4p" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.036349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.047612 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64d55221-60c5-4521-ac84-beed4e9b3993-cert\") pod \"infra-operator-controller-manager-54dc5b8f8d-lp7kj\" (UID: \"64d55221-60c5-4521-ac84-beed4e9b3993\") " pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.170773 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.340023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.348129 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f915558a-c0ab-4d8e-a427-e56a7588e382-cert\") pod \"openstack-baremetal-operator-controller-manager-6f7958d774rr6kf\" (UID: \"f915558a-c0ab-4d8e-a427-e56a7588e382\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.403628 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj"] Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.501948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" event={"ID":"64d55221-60c5-4521-ac84-beed4e9b3993","Type":"ContainerStarted","Data":"0be6213c7eab4cfb969c5684d05680002c055fb080ed20ba4cb83fbda2a9bdb9"} Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.537394 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.789276 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf"] Mar 13 14:24:38 crc kubenswrapper[4907]: W0313 14:24:38.801963 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf915558a_c0ab_4d8e_a427_e56a7588e382.slice/crio-75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe WatchSource:0}: Error finding container 75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe: Status 404 returned error can't find the container with id 75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.957461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.957497 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.964224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-webhook-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:38 crc kubenswrapper[4907]: I0313 14:24:38.964982 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9ac47edf-03ab-4e93-b238-2aa00765ab06-metrics-certs\") pod \"openstack-operator-controller-manager-6484b7b757-qldxp\" (UID: \"9ac47edf-03ab-4e93-b238-2aa00765ab06\") " pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:39 crc kubenswrapper[4907]: I0313 14:24:39.045369 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:39 crc kubenswrapper[4907]: I0313 14:24:39.490415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp"] Mar 13 14:24:39 crc kubenswrapper[4907]: I0313 14:24:39.508021 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" event={"ID":"f915558a-c0ab-4d8e-a427-e56a7588e382","Type":"ContainerStarted","Data":"75c38ddcdad2111a6bf9a09fd763d9ba0db0b2f0bd3aeb1f22fdd968692242fe"} Mar 13 14:24:41 crc kubenswrapper[4907]: W0313 14:24:41.691823 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ac47edf_03ab_4e93_b238_2aa00765ab06.slice/crio-4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1 WatchSource:0}: Error finding container 4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1: Status 404 returned error can't find the container with id 4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1 Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.530562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" event={"ID":"e8134e91-4b9d-4e3c-9aec-eba18acabdce","Type":"ContainerStarted","Data":"bedb2a7348ed35c5b50d4408da49c01fe6c8c250dacc06b7cb57beaed77557f4"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.531060 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.531852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" event={"ID":"4ab8f569-de56-4fea-9be0-07908063522a","Type":"ContainerStarted","Data":"7cfcdad4ee4313cbae8d6e85cbba93863fe24e32bb46b602f6c5107a517a932b"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.532107 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.532906 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" event={"ID":"9ac47edf-03ab-4e93-b238-2aa00765ab06","Type":"ContainerStarted","Data":"e00335a5fbe716a9217486cf5b472f2201008b6b99dbf7c88c2718632c7cd36a"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.532954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" event={"ID":"9ac47edf-03ab-4e93-b238-2aa00765ab06","Type":"ContainerStarted","Data":"4dbfcfd327b82f91e60a91bb090ca1b38f55a0c6e523350af485793c5f2a5ed1"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.533016 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.534248 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" event={"ID":"29201cb7-c056-4e4e-b0da-cc775c7ef11f","Type":"ContainerStarted","Data":"74583fc9170a3e56782dd10b6fbd0dbdb6d034e5d6a17b6fb0a275c783ef4c8e"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.534427 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.536370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" event={"ID":"c55515f1-509e-4a4e-813d-fbc466fca489","Type":"ContainerStarted","Data":"d2c7db2e6567e3827e93800958d2e7f4cf9db08c46c28bc01c5ce56ac4b44b3d"} Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.536654 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.557606 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" podStartSLOduration=2.423232236 podStartE2EDuration="36.557589066s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.664877628 +0000 UTC m=+1146.564665317" lastFinishedPulling="2026-03-13 14:24:41.799234468 +0000 UTC m=+1180.699022147" observedRunningTime="2026-03-13 14:24:42.554282245 +0000 UTC m=+1181.454069934" watchObservedRunningTime="2026-03-13 14:24:42.557589066 +0000 UTC m=+1181.457376755" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.579355 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" podStartSLOduration=2.713793994 podStartE2EDuration="36.579340353s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.860966812 +0000 UTC m=+1146.760754501" lastFinishedPulling="2026-03-13 14:24:41.726513171 +0000 UTC m=+1180.626300860" observedRunningTime="2026-03-13 14:24:42.574960261 +0000 UTC m=+1181.474747950" watchObservedRunningTime="2026-03-13 14:24:42.579340353 +0000 UTC m=+1181.479128032" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.603872 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" podStartSLOduration=2.396819621 podStartE2EDuration="36.603857585s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.693148966 +0000 UTC m=+1146.592936655" lastFinishedPulling="2026-03-13 14:24:41.90018693 +0000 UTC m=+1180.799974619" observedRunningTime="2026-03-13 14:24:42.600095731 +0000 UTC m=+1181.499883420" watchObservedRunningTime="2026-03-13 14:24:42.603857585 +0000 UTC m=+1181.503645274" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.616700 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" podStartSLOduration=2.74386588 podStartE2EDuration="36.616681023s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:07.859420278 +0000 UTC m=+1146.759207967" lastFinishedPulling="2026-03-13 14:24:41.732235411 +0000 UTC m=+1180.632023110" observedRunningTime="2026-03-13 14:24:42.614834512 +0000 UTC m=+1181.514622201" watchObservedRunningTime="2026-03-13 14:24:42.616681023 +0000 UTC m=+1181.516468712" Mar 13 14:24:42 crc kubenswrapper[4907]: I0313 14:24:42.654988 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" podStartSLOduration=36.65497235 podStartE2EDuration="36.65497235s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:24:42.653872489 +0000 UTC m=+1181.553660178" watchObservedRunningTime="2026-03-13 14:24:42.65497235 +0000 UTC m=+1181.554760039" Mar 13 14:24:44 crc kubenswrapper[4907]: I0313 14:24:44.562120 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" event={"ID":"f915558a-c0ab-4d8e-a427-e56a7588e382","Type":"ContainerStarted","Data":"c490d1d54cec4cc6b65cac26ba1a01b62916a5722f82287e8f4992cf2722d681"} Mar 13 14:24:44 crc kubenswrapper[4907]: I0313 14:24:44.562495 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:44 crc kubenswrapper[4907]: I0313 14:24:44.600796 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" podStartSLOduration=33.758992032 podStartE2EDuration="38.600777705s" podCreationTimestamp="2026-03-13 14:24:06 +0000 UTC" firstStartedPulling="2026-03-13 14:24:38.804011993 +0000 UTC m=+1177.703799692" lastFinishedPulling="2026-03-13 14:24:43.645797676 +0000 UTC m=+1182.545585365" observedRunningTime="2026-03-13 14:24:44.593530922 +0000 UTC m=+1183.493318621" watchObservedRunningTime="2026-03-13 14:24:44.600777705 +0000 UTC m=+1183.500565394" Mar 13 14:24:45 crc kubenswrapper[4907]: I0313 14:24:45.571921 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" event={"ID":"64d55221-60c5-4521-ac84-beed4e9b3993","Type":"ContainerStarted","Data":"1a0c55227e3c894cc4a86677ca0790696d229c5bb1f62e53708a5514a6737c05"} Mar 13 14:24:45 crc kubenswrapper[4907]: I0313 14:24:45.572316 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:24:45 crc kubenswrapper[4907]: I0313 14:24:45.588355 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" podStartSLOduration=34.18932223 podStartE2EDuration="40.58832268s" podCreationTimestamp="2026-03-13 14:24:05 +0000 UTC" firstStartedPulling="2026-03-13 14:24:38.412107225 +0000 UTC m=+1177.311894914" lastFinishedPulling="2026-03-13 14:24:44.811107675 +0000 UTC m=+1183.710895364" observedRunningTime="2026-03-13 14:24:45.587506327 +0000 UTC m=+1184.487294016" watchObservedRunningTime="2026-03-13 14:24:45.58832268 +0000 UTC m=+1184.488110369" Mar 13 14:24:48 crc kubenswrapper[4907]: I0313 14:24:48.041518 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:24:48 crc kubenswrapper[4907]: I0313 14:24:48.041599 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:24:48 crc kubenswrapper[4907]: I0313 14:24:48.545389 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f7958d774rr6kf" Mar 13 14:24:49 crc kubenswrapper[4907]: I0313 14:24:49.050932 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6484b7b757-qldxp" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.597400 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-57b484b4df-m8j7m" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.643240 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-gvgdl" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.643734 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw" Mar 13 14:24:56 crc kubenswrapper[4907]: I0313 14:24:56.681757 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7f84474648-ql4kd" Mar 13 14:24:58 crc kubenswrapper[4907]: I0313 14:24:58.176404 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-54dc5b8f8d-lp7kj" Mar 13 14:25:02 crc kubenswrapper[4907]: I0313 14:25:02.688567 4907 scope.go:117] "RemoveContainer" containerID="7377ca75e3097f224f1949f7536a8f7a1c19b44224ea656c9d4dd6e3d82e1652" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.888627 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.891283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894189 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894541 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894536 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.894615 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-wgpxh" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.895769 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.895835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.906170 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.961226 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.962619 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.968831 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:13 crc kubenswrapper[4907]: I0313 14:25:13.969734 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.005693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.005746 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.006806 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.046553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"dnsmasq-dns-5448ff6dc7-pgq4q\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.107497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.107635 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.107707 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.207953 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.208603 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.208669 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.208704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.209434 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.209535 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.243773 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"dnsmasq-dns-64696987c5-gdk7n\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.276575 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.588303 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:14 crc kubenswrapper[4907]: W0313 14:25:14.691033 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod152380e7_1c7a_4c89_a14e_dd179009f6af.slice/crio-5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b WatchSource:0}: Error finding container 5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b: Status 404 returned error can't find the container with id 5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.692828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.789421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" event={"ID":"874829e6-8bd5-4040-a493-317a2af4fa35","Type":"ContainerStarted","Data":"d7f24309913163b7f0ac4df94b2d551597928123506fa9a0c942b473d75c1940"} Mar 13 14:25:14 crc kubenswrapper[4907]: I0313 14:25:14.790512 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" event={"ID":"152380e7-1c7a-4c89-a14e-dd179009f6af","Type":"ContainerStarted","Data":"5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b"} Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.735637 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.776671 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.778289 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.793016 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.960829 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.960971 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:16 crc kubenswrapper[4907]: I0313 14:25:16.961098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.062250 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.062399 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.062430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.063542 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.064069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.105108 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"dnsmasq-dns-658f55c9f5-dp8jf\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.108468 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.180357 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.227460 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.228532 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.245922 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.370933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.370989 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.371035 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.478372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.478431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.478466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.479951 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.480496 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.498903 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"dnsmasq-dns-54b5dffb47-wttzd\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.551267 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.652772 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.825476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerStarted","Data":"00577cc935dfba6c223104c8f045d104f6284aba771cb8fe435ed068373d1547"} Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.948232 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.949281 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.954740 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.955021 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.955168 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.955326 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-swzzq" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.956665 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.958076 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.958250 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 14:25:17 crc kubenswrapper[4907]: I0313 14:25:17.960402 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.041651 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.041703 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.041749 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.042304 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.042356 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736" gracePeriod=600 Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087680 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087710 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087733 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087765 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087782 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087807 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087825 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087859 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.087894 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.091740 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.188858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189148 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189180 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189219 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189237 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189286 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189304 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189362 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.189395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.190180 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.190655 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.190693 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.191204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.191689 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.191804 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.196529 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.209626 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.209790 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.212846 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.213440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.226498 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.274188 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.347579 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.354904 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360206 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360428 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-msttx" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360615 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.360765 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.361109 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.361252 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.361397 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.370623 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496039 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496100 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496133 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496222 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496283 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496310 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496341 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496361 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.496401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597749 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597787 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597821 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597927 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597946 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.597970 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.598003 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.598023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599021 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599121 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599155 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.599660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603416 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.603561 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.608870 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.613809 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.616911 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.617487 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.694734 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.844304 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736" exitCode=0 Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.844346 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736"} Mar 13 14:25:18 crc kubenswrapper[4907]: I0313 14:25:18.844378 4907 scope.go:117] "RemoveContainer" containerID="abe5ce1bc0297c29b962bd8d44bae1a474036514922a5df8e114b5dafeb79cce" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.415834 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.417278 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.420430 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.420862 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-tptn8" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.421725 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.421863 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.425947 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.445677 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512452 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512510 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512612 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512671 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.512721 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614815 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614839 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614944 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.614972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.616101 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.617411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.619597 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.619671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.620208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.628459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.646748 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.647550 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.655864 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " pod="openstack/openstack-galera-0" Mar 13 14:25:19 crc kubenswrapper[4907]: I0313 14:25:19.746444 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.832832 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.834355 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.835991 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.837777 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.838091 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.838213 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-szj6q" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.843041 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.931773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932266 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932332 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932382 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932409 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.932509 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.953571 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.955741 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.959024 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.959384 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-zr52r" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.959624 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 13 14:25:20 crc kubenswrapper[4907]: I0313 14:25:20.966507 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034057 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034142 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034160 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034271 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034359 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.034376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.035202 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.036238 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.036754 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.036843 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.037716 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.041707 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.056012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.057019 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.058515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140200 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140276 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140303 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.140388 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.142943 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.142946 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.143438 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.145635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.155808 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.162502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"memcached-0\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.271710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:25:21 crc kubenswrapper[4907]: W0313 14:25:21.428686 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod779b1958_f7ed_4c53_a4ff_a6c4a803bebf.slice/crio-54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1 WatchSource:0}: Error finding container 54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1: Status 404 returned error can't find the container with id 54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1 Mar 13 14:25:21 crc kubenswrapper[4907]: I0313 14:25:21.908081 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerStarted","Data":"54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1"} Mar 13 14:25:22 crc kubenswrapper[4907]: I0313 14:25:22.989499 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:25:22 crc kubenswrapper[4907]: I0313 14:25:22.990691 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.000682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.046417 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-48tmv" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.077129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"kube-state-metrics-0\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.178099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"kube-state-metrics-0\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.206939 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"kube-state-metrics-0\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " pod="openstack/kube-state-metrics-0" Mar 13 14:25:23 crc kubenswrapper[4907]: I0313 14:25:23.363653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:25:25 crc kubenswrapper[4907]: I0313 14:25:25.110579 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.516595 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.517753 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.520398 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.520744 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tsp64" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.521836 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.521961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.522360 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.542253 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644596 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644626 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644660 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644679 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.644920 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.645019 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.748866 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749205 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749232 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749270 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749310 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749386 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.749411 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.750666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.750713 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.750965 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.751326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.757003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.757595 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.761303 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.772172 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.772435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-nb-0\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:26 crc kubenswrapper[4907]: I0313 14:25:26.837343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.099127 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.100163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.105247 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-4m476" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.105470 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.105625 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.109297 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.110830 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.116205 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.122395 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155348 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155382 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155476 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155587 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155608 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.155653 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257594 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257639 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257675 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257740 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257785 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257912 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.257963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.258892 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261238 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261323 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261784 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.261931 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.262320 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.262372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.262480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.264379 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.264944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.270055 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.286324 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"ovn-controller-bw2xb\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.289651 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"ovn-controller-ovs-vp6f5\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.426265 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:27 crc kubenswrapper[4907]: I0313 14:25:27.432397 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.818624 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.820072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822077 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822142 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-k4qll" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822577 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.822622 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.831342 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897580 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897800 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897821 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.897987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:29 crc kubenswrapper[4907]: I0313 14:25:29.898055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.998962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999480 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999506 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999557 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:29.999583 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.000120 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.000360 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.000624 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.007124 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.015980 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.025681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.029304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.029487 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.053282 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-sb-0\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.153842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.391719 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:25:30 crc kubenswrapper[4907]: I0313 14:25:30.499411 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.044737 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerStarted","Data":"0dddf7e485bfeda5a825f940bee1a34df372878429700a5d41310e0d0412399b"} Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.045710 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerStarted","Data":"be11df2bf36396e7ed2d074eed97b3e79df7e43e0e8a1a8dec1ef8739586a724"} Mar 13 14:25:31 crc kubenswrapper[4907]: W0313 14:25:31.058867 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4411bd2_c555_433a_9015_f623948b1401.slice/crio-da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8 WatchSource:0}: Error finding container da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8: Status 404 returned error can't find the container with id da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8 Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.111627 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.111866 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qxpcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5448ff6dc7-pgq4q_openstack(152380e7-1c7a-4c89-a14e-dd179009f6af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.115359 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" podUID="152380e7-1c7a-4c89-a14e-dd179009f6af" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.318009 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.318695 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:fbb5be29e9e4fa11f0743e7f74f2e80dcc7445d24770709ea0e038147f752c51,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-br82q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-64696987c5-gdk7n_openstack(874829e6-8bd5-4040-a493-317a2af4fa35): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:25:31 crc kubenswrapper[4907]: E0313 14:25:31.321338 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" podUID="874829e6-8bd5-4040-a493-317a2af4fa35" Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.510213 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.571193 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.587805 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.888098 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:25:31 crc kubenswrapper[4907]: W0313 14:25:31.953756 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f8119b2_e38a_494d_967f_5198b83512c7.slice/crio-4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6 WatchSource:0}: Error finding container 4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6: Status 404 returned error can't find the container with id 4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6 Mar 13 14:25:31 crc kubenswrapper[4907]: I0313 14:25:31.958247 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:25:32 crc kubenswrapper[4907]: W0313 14:25:32.007774 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod056dd756_0d7b_471b_9929_f622d05ad606.slice/crio-8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee WatchSource:0}: Error finding container 8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee: Status 404 returned error can't find the container with id 8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.011879 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.088061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerStarted","Data":"3893671130f88340bc9f4b2c1c7ae2e64ebc3c35392832b178f2caf4f860f9fd"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.090267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerStarted","Data":"84aa3724ad3336159381b40fcc1f3bf130caa7609c9affeddec48f79002eb26c"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.093464 4907 generic.go:334] "Generic (PLEG): container finished" podID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerID="69b8616a20bf2d26e83bca32550971f5d3206d2a8a1805adafdfe85310db55b3" exitCode=0 Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.093547 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerDied","Data":"69b8616a20bf2d26e83bca32550971f5d3206d2a8a1805adafdfe85310db55b3"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.100807 4907 generic.go:334] "Generic (PLEG): container finished" podID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" exitCode=0 Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.105047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerDied","Data":"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.110220 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.116020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.125171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerStarted","Data":"da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8"} Mar 13 14:25:32 crc kubenswrapper[4907]: W0313 14:25:32.131065 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2938e39_f1d8_4cdc_a32c_5d57b8f2034f.slice/crio-c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e WatchSource:0}: Error finding container c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e: Status 404 returned error can't find the container with id c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.136948 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerStarted","Data":"c174ea292d3346503480cf55d66ad04b5cc26c89fac549263ee565359ea5a081"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.139185 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerStarted","Data":"4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.141434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerStarted","Data":"8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.147091 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerStarted","Data":"99abe9d38052154162343b52b5e60bd6560b49e824682c932c7a2a64adbb5d61"} Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.502195 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.570497 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.579456 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") pod \"152380e7-1c7a-4c89-a14e-dd179009f6af\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.579924 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") pod \"152380e7-1c7a-4c89-a14e-dd179009f6af\" (UID: \"152380e7-1c7a-4c89-a14e-dd179009f6af\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.579979 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config" (OuterVolumeSpecName: "config") pod "152380e7-1c7a-4c89-a14e-dd179009f6af" (UID: "152380e7-1c7a-4c89-a14e-dd179009f6af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.580703 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/152380e7-1c7a-4c89-a14e-dd179009f6af-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.589071 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr" (OuterVolumeSpecName: "kube-api-access-qxpcr") pod "152380e7-1c7a-4c89-a14e-dd179009f6af" (UID: "152380e7-1c7a-4c89-a14e-dd179009f6af"). InnerVolumeSpecName "kube-api-access-qxpcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.681462 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") pod \"874829e6-8bd5-4040-a493-317a2af4fa35\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.681551 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") pod \"874829e6-8bd5-4040-a493-317a2af4fa35\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.681608 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") pod \"874829e6-8bd5-4040-a493-317a2af4fa35\" (UID: \"874829e6-8bd5-4040-a493-317a2af4fa35\") " Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.682274 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxpcr\" (UniqueName: \"kubernetes.io/projected/152380e7-1c7a-4c89-a14e-dd179009f6af-kube-api-access-qxpcr\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.682278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "874829e6-8bd5-4040-a493-317a2af4fa35" (UID: "874829e6-8bd5-4040-a493-317a2af4fa35"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.682306 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config" (OuterVolumeSpecName: "config") pod "874829e6-8bd5-4040-a493-317a2af4fa35" (UID: "874829e6-8bd5-4040-a493-317a2af4fa35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.685463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q" (OuterVolumeSpecName: "kube-api-access-br82q") pod "874829e6-8bd5-4040-a493-317a2af4fa35" (UID: "874829e6-8bd5-4040-a493-317a2af4fa35"). InnerVolumeSpecName "kube-api-access-br82q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.784177 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br82q\" (UniqueName: \"kubernetes.io/projected/874829e6-8bd5-4040-a493-317a2af4fa35-kube-api-access-br82q\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.784406 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:32 crc kubenswrapper[4907]: I0313 14:25:32.784416 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/874829e6-8bd5-4040-a493-317a2af4fa35-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.162299 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.162294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5448ff6dc7-pgq4q" event={"ID":"152380e7-1c7a-4c89-a14e-dd179009f6af","Type":"ContainerDied","Data":"5fce8f6cb27b89ad9b7731f40d4a1c6f21ae57078b5536d25290f96168055e8b"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.165931 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerStarted","Data":"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.167370 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerStarted","Data":"c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.168799 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" event={"ID":"874829e6-8bd5-4040-a493-317a2af4fa35","Type":"ContainerDied","Data":"d7f24309913163b7f0ac4df94b2d551597928123506fa9a0c942b473d75c1940"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.168830 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64696987c5-gdk7n" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.171842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerStarted","Data":"9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e"} Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.187867 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" podStartSLOduration=6.205187355 podStartE2EDuration="16.187837705s" podCreationTimestamp="2026-03-13 14:25:17 +0000 UTC" firstStartedPulling="2026-03-13 14:25:21.437749451 +0000 UTC m=+1220.337537140" lastFinishedPulling="2026-03-13 14:25:31.420399801 +0000 UTC m=+1230.320187490" observedRunningTime="2026-03-13 14:25:33.18586266 +0000 UTC m=+1232.085650359" watchObservedRunningTime="2026-03-13 14:25:33.187837705 +0000 UTC m=+1232.087625394" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.211313 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" podStartSLOduration=3.465015686 podStartE2EDuration="17.211292199s" podCreationTimestamp="2026-03-13 14:25:16 +0000 UTC" firstStartedPulling="2026-03-13 14:25:17.664366596 +0000 UTC m=+1216.564154285" lastFinishedPulling="2026-03-13 14:25:31.410643119 +0000 UTC m=+1230.310430798" observedRunningTime="2026-03-13 14:25:33.203688886 +0000 UTC m=+1232.103476595" watchObservedRunningTime="2026-03-13 14:25:33.211292199 +0000 UTC m=+1232.111079888" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.270295 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.284703 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64696987c5-gdk7n"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.299284 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.305708 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5448ff6dc7-pgq4q"] Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.810157 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="152380e7-1c7a-4c89-a14e-dd179009f6af" path="/var/lib/kubelet/pods/152380e7-1c7a-4c89-a14e-dd179009f6af/volumes" Mar 13 14:25:33 crc kubenswrapper[4907]: I0313 14:25:33.810591 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="874829e6-8bd5-4040-a493-317a2af4fa35" path="/var/lib/kubelet/pods/874829e6-8bd5-4040-a493-317a2af4fa35/volumes" Mar 13 14:25:34 crc kubenswrapper[4907]: I0313 14:25:34.178959 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:34 crc kubenswrapper[4907]: I0313 14:25:34.179005 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.111538 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.553101 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.621790 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:37 crc kubenswrapper[4907]: I0313 14:25:37.622214 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" containerID="cri-o://9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e" gracePeriod=10 Mar 13 14:25:38 crc kubenswrapper[4907]: I0313 14:25:38.216003 4907 generic.go:334] "Generic (PLEG): container finished" podID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerID="9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e" exitCode=0 Mar 13 14:25:38 crc kubenswrapper[4907]: I0313 14:25:38.216050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerDied","Data":"9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e"} Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.216612 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.278422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" event={"ID":"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0","Type":"ContainerDied","Data":"00577cc935dfba6c223104c8f045d104f6284aba771cb8fe435ed068373d1547"} Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.278466 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f55c9f5-dp8jf" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.278476 4907 scope.go:117] "RemoveContainer" containerID="9cfe7128fc9a7f1b95d9545bb2adc26bb8cdaafa4e69d6b740d1e1ef7d29828e" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.311909 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") pod \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.312338 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") pod \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.312478 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") pod \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\" (UID: \"b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0\") " Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.316202 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn" (OuterVolumeSpecName: "kube-api-access-684gn") pod "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" (UID: "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0"). InnerVolumeSpecName "kube-api-access-684gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.361130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config" (OuterVolumeSpecName: "config") pod "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" (UID: "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.369234 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" (UID: "b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.413600 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.413627 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.413637 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-684gn\" (UniqueName: \"kubernetes.io/projected/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0-kube-api-access-684gn\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.545639 4907 scope.go:117] "RemoveContainer" containerID="69b8616a20bf2d26e83bca32550971f5d3206d2a8a1805adafdfe85310db55b3" Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.613620 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:40 crc kubenswrapper[4907]: I0313 14:25:40.623586 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-658f55c9f5-dp8jf"] Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.314743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerStarted","Data":"432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.316484 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerStarted","Data":"1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.321865 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerStarted","Data":"65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.322465 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.324712 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerStarted","Data":"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86"} Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.386725 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.087687758 podStartE2EDuration="21.38670885s" podCreationTimestamp="2026-03-13 14:25:20 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.523425281 +0000 UTC m=+1230.423212980" lastFinishedPulling="2026-03-13 14:25:39.822446383 +0000 UTC m=+1238.722234072" observedRunningTime="2026-03-13 14:25:41.37977035 +0000 UTC m=+1240.279558039" watchObservedRunningTime="2026-03-13 14:25:41.38670885 +0000 UTC m=+1240.286496539" Mar 13 14:25:41 crc kubenswrapper[4907]: I0313 14:25:41.806875 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" path="/var/lib/kubelet/pods/b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0/volumes" Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.340109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerStarted","Data":"df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.348160 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerStarted","Data":"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.354566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerStarted","Data":"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.356976 4907 generic.go:334] "Generic (PLEG): container finished" podID="c1b70392-1240-40d9-8128-e7abe29c8398" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" exitCode=0 Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.357023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.368660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerStarted","Data":"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c"} Mar 13 14:25:42 crc kubenswrapper[4907]: I0313 14:25:42.487985 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.968026545 podStartE2EDuration="20.487961632s" podCreationTimestamp="2026-03-13 14:25:22 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.605871698 +0000 UTC m=+1230.505659387" lastFinishedPulling="2026-03-13 14:25:41.125806785 +0000 UTC m=+1240.025594474" observedRunningTime="2026-03-13 14:25:42.474832922 +0000 UTC m=+1241.374620611" watchObservedRunningTime="2026-03-13 14:25:42.487961632 +0000 UTC m=+1241.387749321" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.364604 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.378898 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerStarted","Data":"987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38"} Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.383968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerStarted","Data":"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f"} Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.384024 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerStarted","Data":"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574"} Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.384572 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.384756 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.435789 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-bw2xb" podStartSLOduration=7.898204665 podStartE2EDuration="16.435768591s" podCreationTimestamp="2026-03-13 14:25:27 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.990189186 +0000 UTC m=+1230.889976875" lastFinishedPulling="2026-03-13 14:25:40.527753112 +0000 UTC m=+1239.427540801" observedRunningTime="2026-03-13 14:25:43.431262247 +0000 UTC m=+1242.331049936" watchObservedRunningTime="2026-03-13 14:25:43.435768591 +0000 UTC m=+1242.335556280" Mar 13 14:25:43 crc kubenswrapper[4907]: I0313 14:25:43.449443 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-vp6f5" podStartSLOduration=8.118672503 podStartE2EDuration="16.449426726s" podCreationTimestamp="2026-03-13 14:25:27 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.898159222 +0000 UTC m=+1230.797946911" lastFinishedPulling="2026-03-13 14:25:40.228913435 +0000 UTC m=+1239.128701134" observedRunningTime="2026-03-13 14:25:43.448239533 +0000 UTC m=+1242.348027232" watchObservedRunningTime="2026-03-13 14:25:43.449426726 +0000 UTC m=+1242.349214415" Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.400598 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerStarted","Data":"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb"} Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.404294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerStarted","Data":"1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6"} Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.423474 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.513472542 podStartE2EDuration="20.423452586s" podCreationTimestamp="2026-03-13 14:25:25 +0000 UTC" firstStartedPulling="2026-03-13 14:25:32.033185113 +0000 UTC m=+1230.932972802" lastFinishedPulling="2026-03-13 14:25:44.943165157 +0000 UTC m=+1243.842952846" observedRunningTime="2026-03-13 14:25:45.420465834 +0000 UTC m=+1244.320253533" watchObservedRunningTime="2026-03-13 14:25:45.423452586 +0000 UTC m=+1244.323240275" Mar 13 14:25:45 crc kubenswrapper[4907]: I0313 14:25:45.442318 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.623879986 podStartE2EDuration="17.442298613s" podCreationTimestamp="2026-03-13 14:25:28 +0000 UTC" firstStartedPulling="2026-03-13 14:25:32.137002996 +0000 UTC m=+1231.036790685" lastFinishedPulling="2026-03-13 14:25:44.955421623 +0000 UTC m=+1243.855209312" observedRunningTime="2026-03-13 14:25:45.437546603 +0000 UTC m=+1244.337334302" watchObservedRunningTime="2026-03-13 14:25:45.442298613 +0000 UTC m=+1244.342086302" Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.273704 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.413413 4907 generic.go:334] "Generic (PLEG): container finished" podID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" exitCode=0 Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.413481 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerDied","Data":"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86"} Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.415461 4907 generic.go:334] "Generic (PLEG): container finished" podID="c4411bd2-c555-433a-9015-f623948b1401" containerID="1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729" exitCode=0 Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.415494 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerDied","Data":"1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729"} Mar 13 14:25:46 crc kubenswrapper[4907]: I0313 14:25:46.842694 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.424848 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerStarted","Data":"b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa"} Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.426347 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-bw2xb" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.427083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerStarted","Data":"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a"} Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.498490 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.146814831 podStartE2EDuration="29.49846879s" podCreationTimestamp="2026-03-13 14:25:18 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.077097965 +0000 UTC m=+1229.976885654" lastFinishedPulling="2026-03-13 14:25:40.428751924 +0000 UTC m=+1239.328539613" observedRunningTime="2026-03-13 14:25:47.455842029 +0000 UTC m=+1246.355629718" watchObservedRunningTime="2026-03-13 14:25:47.49846879 +0000 UTC m=+1246.398256479" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.503574 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.43645419 podStartE2EDuration="28.503557659s" podCreationTimestamp="2026-03-13 14:25:19 +0000 UTC" firstStartedPulling="2026-03-13 14:25:30.069412459 +0000 UTC m=+1228.969200148" lastFinishedPulling="2026-03-13 14:25:40.136515928 +0000 UTC m=+1239.036303617" observedRunningTime="2026-03-13 14:25:47.503129767 +0000 UTC m=+1246.402917456" watchObservedRunningTime="2026-03-13 14:25:47.503557659 +0000 UTC m=+1246.403345348" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.837778 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:47 crc kubenswrapper[4907]: I0313 14:25:47.880924 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.155583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.200973 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.435277 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.478309 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.495333 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736291 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:48 crc kubenswrapper[4907]: E0313 14:25:48.736641 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736654 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" Mar 13 14:25:48 crc kubenswrapper[4907]: E0313 14:25:48.736673 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="init" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736679 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="init" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.736913 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b921f7f7-cd29-4ac8-82e5-c69a8f9e2bc0" containerName="dnsmasq-dns" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.737739 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.740234 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.750942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.788129 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.789507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.792562 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.858234 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877730 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877782 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877878 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877939 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.877990 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.911277 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.912492 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.920605 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.920835 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.920997 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-c5q7n" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.921196 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.937797 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:48 crc kubenswrapper[4907]: E0313 14:25:48.938507 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-z85pw ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" podUID="5e9f0d4f-790a-444e-9090-d9c6fa952c2b" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.952567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981850 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981875 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981944 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981966 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.981986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.982359 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.983258 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.983461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.984337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.985520 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.989049 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.989567 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.989700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.990226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.991113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:48 crc kubenswrapper[4907]: I0313 14:25:48.992427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.008755 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"dnsmasq-dns-7988f9db49-2mh7s\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.008762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"ovn-controller-metrics-nhnkq\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.024903 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.082975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083185 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083535 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083700 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083856 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.083988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.084069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.117238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185715 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185920 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.185948 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186005 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186030 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186104 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186142 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.186170 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187089 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187832 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.187972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.188210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.188478 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.188553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.191264 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.192539 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.206727 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.210877 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"ovn-northd-0\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.229610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"dnsmasq-dns-5d944d7b75-sk7j6\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.250452 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.386985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.442615 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.464819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.590748 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.590897 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591040 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") pod \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\" (UID: \"5e9f0d4f-790a-444e-9090-d9c6fa952c2b\") " Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591526 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591526 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config" (OuterVolumeSpecName: "config") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.591776 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.592110 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.592481 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.596073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw" (OuterVolumeSpecName: "kube-api-access-z85pw") pod "5e9f0d4f-790a-444e-9090-d9c6fa952c2b" (UID: "5e9f0d4f-790a-444e-9090-d9c6fa952c2b"). InnerVolumeSpecName "kube-api-access-z85pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.609625 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.694122 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z85pw\" (UniqueName: \"kubernetes.io/projected/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-kube-api-access-z85pw\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.694229 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5e9f0d4f-790a-444e-9090-d9c6fa952c2b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.747171 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.747215 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.833775 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:25:49 crc kubenswrapper[4907]: I0313 14:25:49.883673 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:49 crc kubenswrapper[4907]: W0313 14:25:49.885200 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26827945_75f0_4867_ba04_31ff6428e06a.slice/crio-59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa WatchSource:0}: Error finding container 59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa: Status 404 returned error can't find the container with id 59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.450938 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nhnkq" event={"ID":"37978cc1-fcba-4032-a8b1-6632b61692ff","Type":"ContainerStarted","Data":"5a1bb100c70a1f42d543b5e88b530ca72ed03e93dea7325eb0c88d0809f831ad"} Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.453308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerStarted","Data":"d38876707ec2c1ee40fc15622432cfe811af1e53b494065c48deee03085ef192"} Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.453558 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerStarted","Data":"59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa"} Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.453447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7988f9db49-2mh7s" Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.513357 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:50 crc kubenswrapper[4907]: I0313 14:25:50.539029 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7988f9db49-2mh7s"] Mar 13 14:25:51 crc kubenswrapper[4907]: I0313 14:25:51.156904 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:51 crc kubenswrapper[4907]: I0313 14:25:51.157244 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:51 crc kubenswrapper[4907]: I0313 14:25:51.791992 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e9f0d4f-790a-444e-9090-d9c6fa952c2b" path="/var/lib/kubelet/pods/5e9f0d4f-790a-444e-9090-d9c6fa952c2b/volumes" Mar 13 14:25:52 crc kubenswrapper[4907]: I0313 14:25:52.466639 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nhnkq" event={"ID":"37978cc1-fcba-4032-a8b1-6632b61692ff","Type":"ContainerStarted","Data":"62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a"} Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.160127 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.192145 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.193305 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.203124 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.262621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263016 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263127 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263236 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.263403 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.364980 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365072 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365096 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.365144 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.366147 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.366595 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.367281 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.367648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.368143 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.401834 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"dnsmasq-dns-7b9fd7d84c-wkdxf\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.475684 4907 generic.go:334] "Generic (PLEG): container finished" podID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" exitCode=0 Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.475823 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerDied","Data":"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee"} Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.492209 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-nhnkq" podStartSLOduration=5.492192908 podStartE2EDuration="5.492192908s" podCreationTimestamp="2026-03-13 14:25:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:53.488807854 +0000 UTC m=+1252.388595543" watchObservedRunningTime="2026-03-13 14:25:53.492192908 +0000 UTC m=+1252.391980597" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.513889 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.913961 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 14:25:53 crc kubenswrapper[4907]: I0313 14:25:53.994434 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.199299 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:25:54 crc kubenswrapper[4907]: W0313 14:25:54.202068 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89cd945f_311f_47d1_982a_641f062e4f57.slice/crio-b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476 WatchSource:0}: Error finding container b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476: Status 404 returned error can't find the container with id b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476 Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.291999 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.302037 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304552 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304593 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mrlx5" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304698 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.304991 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.320557 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481270 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481316 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481345 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.481417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.484383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerStarted","Data":"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.484498 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" containerID="cri-o://0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" gracePeriod=10 Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.484547 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.486122 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerStarted","Data":"59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.486381 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerStarted","Data":"e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.486397 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.488271 4907 generic.go:334] "Generic (PLEG): container finished" podID="89cd945f-311f-47d1-982a-641f062e4f57" containerID="b1318d08ab870fbe30d1438d26843f78b5e12e9bf7ba05a68e44fb095db6996e" exitCode=0 Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.488299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerDied","Data":"b1318d08ab870fbe30d1438d26843f78b5e12e9bf7ba05a68e44fb095db6996e"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.488325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerStarted","Data":"b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476"} Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.511320 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" podStartSLOduration=6.5112991529999995 podStartE2EDuration="6.511299153s" podCreationTimestamp="2026-03-13 14:25:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:54.498623265 +0000 UTC m=+1253.398410974" watchObservedRunningTime="2026-03-13 14:25:54.511299153 +0000 UTC m=+1253.411086842" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.549419 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.499962866 podStartE2EDuration="6.54939769s" podCreationTimestamp="2026-03-13 14:25:48 +0000 UTC" firstStartedPulling="2026-03-13 14:25:49.90000902 +0000 UTC m=+1248.799796709" lastFinishedPulling="2026-03-13 14:25:53.949443844 +0000 UTC m=+1252.849231533" observedRunningTime="2026-03-13 14:25:54.540202977 +0000 UTC m=+1253.439990666" watchObservedRunningTime="2026-03-13 14:25:54.54939769 +0000 UTC m=+1253.449185379" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582491 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582610 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582706 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582762 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.582784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.583489 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: E0313 14:25:54.583595 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:54 crc kubenswrapper[4907]: E0313 14:25:54.583613 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:54 crc kubenswrapper[4907]: E0313 14:25:54.583661 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:25:55.08364227 +0000 UTC m=+1253.983429959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.583713 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.583896 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.589605 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.604671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.606787 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.871025 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.990962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.991230 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") pod \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\" (UID: \"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b\") " Mar 13 14:25:54 crc kubenswrapper[4907]: I0313 14:25:54.997576 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6" (OuterVolumeSpecName: "kube-api-access-7kqs6") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "kube-api-access-7kqs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.032319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.037073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.040938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.043265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config" (OuterVolumeSpecName: "config") pod "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" (UID: "ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093399 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093546 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093563 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093574 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kqs6\" (UniqueName: \"kubernetes.io/projected/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-kube-api-access-7kqs6\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093586 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.093589 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.093617 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.093664 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:25:56.093649736 +0000 UTC m=+1254.993437425 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.093597 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.279416 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.481992 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496299 4907 generic.go:334] "Generic (PLEG): container finished" podID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" exitCode=0 Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496359 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerDied","Data":"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc"} Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496367 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496384 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d944d7b75-sk7j6" event={"ID":"ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b","Type":"ContainerDied","Data":"d38876707ec2c1ee40fc15622432cfe811af1e53b494065c48deee03085ef192"} Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.496402 4907 scope.go:117] "RemoveContainer" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.498775 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerStarted","Data":"ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96"} Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.517589 4907 scope.go:117] "RemoveContainer" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.542010 4907 scope.go:117] "RemoveContainer" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.543860 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc\": container with ID starting with 0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc not found: ID does not exist" containerID="0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.543935 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc"} err="failed to get container status \"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc\": rpc error: code = NotFound desc = could not find container \"0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc\": container with ID starting with 0da5a6353eaf69e9bcfb1c6797b8931d8c2f75f259f947daed552ac9ac517adc not found: ID does not exist" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.543965 4907 scope.go:117] "RemoveContainer" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" Mar 13 14:25:55 crc kubenswrapper[4907]: E0313 14:25:55.547165 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee\": container with ID starting with e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee not found: ID does not exist" containerID="e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.547197 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee"} err="failed to get container status \"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee\": rpc error: code = NotFound desc = could not find container \"e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee\": container with ID starting with e663df1a6e0053a10caca381de5f2223c40515b9e82b5695a1a4d758c5c706ee not found: ID does not exist" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.578919 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" podStartSLOduration=2.578895302 podStartE2EDuration="2.578895302s" podCreationTimestamp="2026-03-13 14:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:55.560094086 +0000 UTC m=+1254.459881775" watchObservedRunningTime="2026-03-13 14:25:55.578895302 +0000 UTC m=+1254.478683011" Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.586966 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.593609 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d944d7b75-sk7j6"] Mar 13 14:25:55 crc kubenswrapper[4907]: I0313 14:25:55.791182 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" path="/var/lib/kubelet/pods/ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b/volumes" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.109758 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.109987 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.110203 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.110279 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:25:58.110255613 +0000 UTC m=+1257.010043342 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.508593 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521073 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.521403 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="init" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521418 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="init" Mar 13 14:25:56 crc kubenswrapper[4907]: E0313 14:25:56.521436 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521443 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.521616 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea92b3fe-e6a1-409a-9d1a-eb6a76844f8b" containerName="dnsmasq-dns" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.522840 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.531105 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.532135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.535873 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.551840 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.561081 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.618696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.618787 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.619023 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.619405 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721412 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.721457 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.722301 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.722413 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.743130 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"glance-17a8-account-create-update-xv8sm\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.749484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"glance-db-create-sgpkl\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.852278 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:25:56 crc kubenswrapper[4907]: I0313 14:25:56.861795 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.278104 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:25:57 crc kubenswrapper[4907]: W0313 14:25:57.280157 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7d2639_4229_4286_be5c_d8b15ed91d17.slice/crio-64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc WatchSource:0}: Error finding container 64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc: Status 404 returned error can't find the container with id 64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.354903 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:25:57 crc kubenswrapper[4907]: W0313 14:25:57.364626 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb6c089_0112_43f4_8731_c2d68932795a.slice/crio-99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652 WatchSource:0}: Error finding container 99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652: Status 404 returned error can't find the container with id 99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652 Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.520688 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerStarted","Data":"3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.520731 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerStarted","Data":"64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.524330 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerStarted","Data":"b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.524396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerStarted","Data":"99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652"} Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.539874 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-sgpkl" podStartSLOduration=1.539851932 podStartE2EDuration="1.539851932s" podCreationTimestamp="2026-03-13 14:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:57.534348452 +0000 UTC m=+1256.434136141" watchObservedRunningTime="2026-03-13 14:25:57.539851932 +0000 UTC m=+1256.439639641" Mar 13 14:25:57 crc kubenswrapper[4907]: I0313 14:25:57.553424 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-17a8-account-create-update-xv8sm" podStartSLOduration=1.553405905 podStartE2EDuration="1.553405905s" podCreationTimestamp="2026-03-13 14:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:57.552416557 +0000 UTC m=+1256.452204236" watchObservedRunningTime="2026-03-13 14:25:57.553405905 +0000 UTC m=+1256.453193594" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.163322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:25:58 crc kubenswrapper[4907]: E0313 14:25:58.163586 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:25:58 crc kubenswrapper[4907]: E0313 14:25:58.163829 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:25:58 crc kubenswrapper[4907]: E0313 14:25:58.163926 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:26:02.163875069 +0000 UTC m=+1261.063662768 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.287069 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.288931 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.296213 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.296296 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.296443 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.309288 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.357805 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.358768 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.361266 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366521 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366571 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366695 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366728 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.366758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.370388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.467974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468029 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468106 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468146 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468165 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.468800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.473243 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.473363 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.474033 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.483447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"swift-ring-rebalance-9fxfh\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.532215 4907 generic.go:334] "Generic (PLEG): container finished" podID="cbb6c089-0112-43f4-8731-c2d68932795a" containerID="b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a" exitCode=0 Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.532278 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerDied","Data":"b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a"} Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.535028 4907 generic.go:334] "Generic (PLEG): container finished" podID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerID="3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff" exitCode=0 Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.535059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerDied","Data":"3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff"} Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.569863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.569971 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.570701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.588092 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"root-account-create-update-5lhdv\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.613331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:25:58 crc kubenswrapper[4907]: I0313 14:25:58.722599 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:25:59 crc kubenswrapper[4907]: W0313 14:25:59.031103 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2585497_6837_4d17_8d51_c3d7879fdb46.slice/crio-c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5 WatchSource:0}: Error finding container c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5: Status 404 returned error can't find the container with id c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5 Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.036395 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.178443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.547157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerStarted","Data":"c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5"} Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.551267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerStarted","Data":"c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226"} Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.551315 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerStarted","Data":"d2489c29658426420882959df704422f457e13151fc0758486e6c026a736b39b"} Mar 13 14:25:59 crc kubenswrapper[4907]: I0313 14:25:59.577148 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-5lhdv" podStartSLOduration=1.577073028 podStartE2EDuration="1.577073028s" podCreationTimestamp="2026-03-13 14:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:25:59.56948906 +0000 UTC m=+1258.469276759" watchObservedRunningTime="2026-03-13 14:25:59.577073028 +0000 UTC m=+1258.476860717" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.133760 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.135032 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.136786 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.136975 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.137909 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.140198 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.201675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"auto-csr-approver-29556866-fwg48\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.210615 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.217731 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303224 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") pod \"4d7d2639-4229-4286-be5c-d8b15ed91d17\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303408 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") pod \"4d7d2639-4229-4286-be5c-d8b15ed91d17\" (UID: \"4d7d2639-4229-4286-be5c-d8b15ed91d17\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303467 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") pod \"cbb6c089-0112-43f4-8731-c2d68932795a\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303499 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") pod \"cbb6c089-0112-43f4-8731-c2d68932795a\" (UID: \"cbb6c089-0112-43f4-8731-c2d68932795a\") " Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.303943 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d7d2639-4229-4286-be5c-d8b15ed91d17" (UID: "4d7d2639-4229-4286-be5c-d8b15ed91d17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304292 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cbb6c089-0112-43f4-8731-c2d68932795a" (UID: "cbb6c089-0112-43f4-8731-c2d68932795a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304459 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"auto-csr-approver-29556866-fwg48\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304540 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d7d2639-4229-4286-be5c-d8b15ed91d17-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.304556 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cbb6c089-0112-43f4-8731-c2d68932795a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.309135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n" (OuterVolumeSpecName: "kube-api-access-zkz6n") pod "cbb6c089-0112-43f4-8731-c2d68932795a" (UID: "cbb6c089-0112-43f4-8731-c2d68932795a"). InnerVolumeSpecName "kube-api-access-zkz6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.309466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf" (OuterVolumeSpecName: "kube-api-access-4q8qf") pod "4d7d2639-4229-4286-be5c-d8b15ed91d17" (UID: "4d7d2639-4229-4286-be5c-d8b15ed91d17"). InnerVolumeSpecName "kube-api-access-4q8qf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.322191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"auto-csr-approver-29556866-fwg48\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.406062 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkz6n\" (UniqueName: \"kubernetes.io/projected/cbb6c089-0112-43f4-8731-c2d68932795a-kube-api-access-zkz6n\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.406098 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q8qf\" (UniqueName: \"kubernetes.io/projected/4d7d2639-4229-4286-be5c-d8b15ed91d17-kube-api-access-4q8qf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.538405 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.568570 4907 generic.go:334] "Generic (PLEG): container finished" podID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerID="c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226" exitCode=0 Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.568667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerDied","Data":"c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226"} Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.574831 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-xv8sm" event={"ID":"cbb6c089-0112-43f4-8731-c2d68932795a","Type":"ContainerDied","Data":"99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652"} Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.574869 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99873ca15b9fb22afcd6ec2e9a14245394d9db6cee5f1f289e90f95b20ea6652" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.575006 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-xv8sm" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.578813 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sgpkl" event={"ID":"4d7d2639-4229-4286-be5c-d8b15ed91d17","Type":"ContainerDied","Data":"64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc"} Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.578840 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sgpkl" Mar 13 14:26:00 crc kubenswrapper[4907]: I0313 14:26:00.578857 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64d9b417d3e9f250fd98eb8db5c1fbf579e78ade9d4cbc4644beb2dad8d75fbc" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.761696 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:26:01 crc kubenswrapper[4907]: E0313 14:26:01.762514 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerName="mariadb-database-create" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762533 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerName="mariadb-database-create" Mar 13 14:26:01 crc kubenswrapper[4907]: E0313 14:26:01.762566 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" containerName="mariadb-account-create-update" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" containerName="mariadb-account-create-update" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762769 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" containerName="mariadb-account-create-update" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.762785 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" containerName="mariadb-database-create" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.763425 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.766029 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.766286 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fc9j2" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.775990 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.827503 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.827874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.828401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.828826 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930301 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930392 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.930593 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.937985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.938074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.940764 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:01 crc kubenswrapper[4907]: I0313 14:26:01.949509 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"glance-db-sync-4v4ps\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.097099 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.212718 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.213782 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.225871 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.236391 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:02 crc kubenswrapper[4907]: E0313 14:26:02.236629 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:26:02 crc kubenswrapper[4907]: E0313 14:26:02.236646 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:26:02 crc kubenswrapper[4907]: E0313 14:26:02.236692 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:26:10.236676855 +0000 UTC m=+1269.136464544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.314967 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.316099 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.319469 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.331738 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.337636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.337746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.439559 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.439628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.439873 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.440034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.440843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.456522 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"keystone-db-create-hg688\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.526410 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.527557 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.532632 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.537196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.542857 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.543011 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.543781 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.543844 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.545079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.547233 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.552315 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.563158 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"keystone-04a8-account-create-update-ws5ht\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.608284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5lhdv" event={"ID":"f12f5638-4cb2-4424-86d8-5c0c829d82d5","Type":"ContainerDied","Data":"d2489c29658426420882959df704422f457e13151fc0758486e6c026a736b39b"} Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.608321 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2489c29658426420882959df704422f457e13151fc0758486e6c026a736b39b" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.634017 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.637006 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644437 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.644518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.745595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") pod \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.745962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") pod \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\" (UID: \"f12f5638-4cb2-4424-86d8-5c0c829d82d5\") " Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746200 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746230 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746271 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f12f5638-4cb2-4424-86d8-5c0c829d82d5" (UID: "f12f5638-4cb2-4424-86d8-5c0c829d82d5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746354 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.746416 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f12f5638-4cb2-4424-86d8-5c0c829d82d5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.747054 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.747383 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.751498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb" (OuterVolumeSpecName: "kube-api-access-zqzfb") pod "f12f5638-4cb2-4424-86d8-5c0c829d82d5" (UID: "f12f5638-4cb2-4424-86d8-5c0c829d82d5"). InnerVolumeSpecName "kube-api-access-zqzfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.767343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"placement-db-create-cnnxs\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.770482 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"placement-9c64-account-create-update-v65tv\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.847728 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqzfb\" (UniqueName: \"kubernetes.io/projected/f12f5638-4cb2-4424-86d8-5c0c829d82d5-kube-api-access-zqzfb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.848045 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:02 crc kubenswrapper[4907]: I0313 14:26:02.864157 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.094721 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.106397 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d930b98_c1af_4e47_a663_86afa9484856.slice/crio-c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9 WatchSource:0}: Error finding container c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9: Status 404 returned error can't find the container with id c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9 Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.171240 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.266053 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.306895 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.310637 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37ba2d41_5de4_48cb_aad6_59df13307ed1.slice/crio-85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f WatchSource:0}: Error finding container 85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f: Status 404 returned error can't find the container with id 85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.387811 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.394640 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.401065 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f2b7163_b5d4_4575_a72c_e6ad1b3137df.slice/crio-03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0 WatchSource:0}: Error finding container 03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0: Status 404 returned error can't find the container with id 03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0 Mar 13 14:26:03 crc kubenswrapper[4907]: W0313 14:26:03.404027 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3db3f0b6_9cae_40fa_b54d_3ed06c568c9c.slice/crio-46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553 WatchSource:0}: Error finding container 46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553: Status 404 returned error can't find the container with id 46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553 Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.516620 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.583310 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.583551 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" containerID="cri-o://923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" gracePeriod=10 Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.621248 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerStarted","Data":"e0469ec6298d86f707976b7020c89f6d29a4be70c9a77f315ceb22e64cbb4824"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.623040 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerStarted","Data":"d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.623134 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerStarted","Data":"03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.627143 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerStarted","Data":"ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.632557 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerStarted","Data":"93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.632596 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerStarted","Data":"485fdae7de5b84d520754700103a7b2d1a2699988fd5bac3363e984c2ccb2b0e"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.634856 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerStarted","Data":"1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.634927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerStarted","Data":"46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.635967 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerStarted","Data":"c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.639568 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5lhdv" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.643466 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerStarted","Data":"3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.643514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerStarted","Data":"85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f"} Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.663231 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-cnnxs" podStartSLOduration=1.66321323 podStartE2EDuration="1.66321323s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.643929541 +0000 UTC m=+1262.543717230" watchObservedRunningTime="2026-03-13 14:26:03.66321323 +0000 UTC m=+1262.563000919" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.675590 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-9fxfh" podStartSLOduration=2.003299873 podStartE2EDuration="5.67557094s" podCreationTimestamp="2026-03-13 14:25:58 +0000 UTC" firstStartedPulling="2026-03-13 14:25:59.040315998 +0000 UTC m=+1257.940103687" lastFinishedPulling="2026-03-13 14:26:02.712587065 +0000 UTC m=+1261.612374754" observedRunningTime="2026-03-13 14:26:03.668424994 +0000 UTC m=+1262.568212683" watchObservedRunningTime="2026-03-13 14:26:03.67557094 +0000 UTC m=+1262.575358629" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.686898 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-9c64-account-create-update-v65tv" podStartSLOduration=1.68687187 podStartE2EDuration="1.68687187s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.684409472 +0000 UTC m=+1262.584197161" watchObservedRunningTime="2026-03-13 14:26:03.68687187 +0000 UTC m=+1262.586659559" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.710453 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-04a8-account-create-update-ws5ht" podStartSLOduration=1.710435197 podStartE2EDuration="1.710435197s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.701661006 +0000 UTC m=+1262.601448695" watchObservedRunningTime="2026-03-13 14:26:03.710435197 +0000 UTC m=+1262.610222886" Mar 13 14:26:03 crc kubenswrapper[4907]: I0313 14:26:03.727851 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-hg688" podStartSLOduration=1.7278339649999999 podStartE2EDuration="1.727833965s" podCreationTimestamp="2026-03-13 14:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:03.724407521 +0000 UTC m=+1262.624195210" watchObservedRunningTime="2026-03-13 14:26:03.727833965 +0000 UTC m=+1262.627621654" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.072596 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.172190 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") pod \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.172247 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") pod \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.172274 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") pod \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\" (UID: \"779b1958-f7ed-4c53-a4ff-a6c4a803bebf\") " Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.177737 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb" (OuterVolumeSpecName: "kube-api-access-bmczb") pod "779b1958-f7ed-4c53-a4ff-a6c4a803bebf" (UID: "779b1958-f7ed-4c53-a4ff-a6c4a803bebf"). InnerVolumeSpecName "kube-api-access-bmczb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.230438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config" (OuterVolumeSpecName: "config") pod "779b1958-f7ed-4c53-a4ff-a6c4a803bebf" (UID: "779b1958-f7ed-4c53-a4ff-a6c4a803bebf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.246378 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "779b1958-f7ed-4c53-a4ff-a6c4a803bebf" (UID: "779b1958-f7ed-4c53-a4ff-a6c4a803bebf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.274582 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.274621 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.274634 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmczb\" (UniqueName: \"kubernetes.io/projected/779b1958-f7ed-4c53-a4ff-a6c4a803bebf-kube-api-access-bmczb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.649923 4907 generic.go:334] "Generic (PLEG): container finished" podID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerID="1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.649973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerDied","Data":"1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.653177 4907 generic.go:334] "Generic (PLEG): container finished" podID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerID="93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.653260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerDied","Data":"93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.654751 4907 generic.go:334] "Generic (PLEG): container finished" podID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerID="3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.654795 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerDied","Data":"3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.656685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerStarted","Data":"37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.663444 4907 generic.go:334] "Generic (PLEG): container finished" podID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerID="d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.663660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerDied","Data":"d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.667919 4907 generic.go:334] "Generic (PLEG): container finished" podID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" exitCode=0 Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.667983 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.668006 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerDied","Data":"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.668044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54b5dffb47-wttzd" event={"ID":"779b1958-f7ed-4c53-a4ff-a6c4a803bebf","Type":"ContainerDied","Data":"54b19297eaef464c5a401c18dcda022f8c38c1cecb9978d5b544939e48c81aa1"} Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.668060 4907 scope.go:117] "RemoveContainer" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.692368 4907 scope.go:117] "RemoveContainer" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.729679 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556866-fwg48" podStartSLOduration=3.871471579 podStartE2EDuration="4.729612175s" podCreationTimestamp="2026-03-13 14:26:00 +0000 UTC" firstStartedPulling="2026-03-13 14:26:03.108751744 +0000 UTC m=+1262.008539433" lastFinishedPulling="2026-03-13 14:26:03.96689234 +0000 UTC m=+1262.866680029" observedRunningTime="2026-03-13 14:26:04.716073774 +0000 UTC m=+1263.615861463" watchObservedRunningTime="2026-03-13 14:26:04.729612175 +0000 UTC m=+1263.629399864" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.742924 4907 scope.go:117] "RemoveContainer" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" Mar 13 14:26:04 crc kubenswrapper[4907]: E0313 14:26:04.746161 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c\": container with ID starting with 923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c not found: ID does not exist" containerID="923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.746403 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c"} err="failed to get container status \"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c\": rpc error: code = NotFound desc = could not find container \"923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c\": container with ID starting with 923c944515f9f62d3a3c73bba3a0205c7dccd6753e7da2435a2be1308a3d0b7c not found: ID does not exist" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.746507 4907 scope.go:117] "RemoveContainer" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" Mar 13 14:26:04 crc kubenswrapper[4907]: E0313 14:26:04.747295 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a\": container with ID starting with cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a not found: ID does not exist" containerID="cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.747341 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a"} err="failed to get container status \"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a\": rpc error: code = NotFound desc = could not find container \"cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a\": container with ID starting with cd0379440a1db9f081229f5637156055839f155515dc1039d28132ecd64ab32a not found: ID does not exist" Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.763678 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.770550 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54b5dffb47-wttzd"] Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.815999 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:26:04 crc kubenswrapper[4907]: I0313 14:26:04.823622 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-5lhdv"] Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.679754 4907 generic.go:334] "Generic (PLEG): container finished" podID="7d930b98-c1af-4e47-a663-86afa9484856" containerID="37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25" exitCode=0 Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.679852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerDied","Data":"37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25"} Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.791356 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" path="/var/lib/kubelet/pods/779b1958-f7ed-4c53-a4ff-a6c4a803bebf/volumes" Mar 13 14:26:05 crc kubenswrapper[4907]: I0313 14:26:05.792397 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" path="/var/lib/kubelet/pods/f12f5638-4cb2-4424-86d8-5c0c829d82d5/volumes" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.032119 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.106282 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") pod \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.106789 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") pod \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\" (UID: \"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.107389 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" (UID: "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.114118 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg" (OuterVolumeSpecName: "kube-api-access-h5wpg") pod "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" (UID: "3db3f0b6-9cae-40fa-b54d-3ed06c568c9c"). InnerVolumeSpecName "kube-api-access-h5wpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.191961 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.197468 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.209345 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.209372 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5wpg\" (UniqueName: \"kubernetes.io/projected/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c-kube-api-access-h5wpg\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.209827 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") pod \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310343 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") pod \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") pod \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\" (UID: \"3f2b7163-b5d4-4575-a72c-e6ad1b3137df\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") pod \"37ba2d41-5de4-48cb-aad6-59df13307ed1\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310535 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") pod \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\" (UID: \"9c985ff9-3545-4544-acc8-ca9eb2bfad6a\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.310569 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") pod \"37ba2d41-5de4-48cb-aad6-59df13307ed1\" (UID: \"37ba2d41-5de4-48cb-aad6-59df13307ed1\") " Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.311178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c985ff9-3545-4544-acc8-ca9eb2bfad6a" (UID: "9c985ff9-3545-4544-acc8-ca9eb2bfad6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.311857 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f2b7163-b5d4-4575-a72c-e6ad1b3137df" (UID: "3f2b7163-b5d4-4575-a72c-e6ad1b3137df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.312099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37ba2d41-5de4-48cb-aad6-59df13307ed1" (UID: "37ba2d41-5de4-48cb-aad6-59df13307ed1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.314113 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l" (OuterVolumeSpecName: "kube-api-access-5cr4l") pod "3f2b7163-b5d4-4575-a72c-e6ad1b3137df" (UID: "3f2b7163-b5d4-4575-a72c-e6ad1b3137df"). InnerVolumeSpecName "kube-api-access-5cr4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.315134 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf" (OuterVolumeSpecName: "kube-api-access-qcmkf") pod "9c985ff9-3545-4544-acc8-ca9eb2bfad6a" (UID: "9c985ff9-3545-4544-acc8-ca9eb2bfad6a"). InnerVolumeSpecName "kube-api-access-qcmkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.315598 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw" (OuterVolumeSpecName: "kube-api-access-b4bzw") pod "37ba2d41-5de4-48cb-aad6-59df13307ed1" (UID: "37ba2d41-5de4-48cb-aad6-59df13307ed1"). InnerVolumeSpecName "kube-api-access-b4bzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412280 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4bzw\" (UniqueName: \"kubernetes.io/projected/37ba2d41-5de4-48cb-aad6-59df13307ed1-kube-api-access-b4bzw\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412318 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcmkf\" (UniqueName: \"kubernetes.io/projected/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-kube-api-access-qcmkf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412327 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37ba2d41-5de4-48cb-aad6-59df13307ed1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412336 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412345 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c985ff9-3545-4544-acc8-ca9eb2bfad6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.412355 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cr4l\" (UniqueName: \"kubernetes.io/projected/3f2b7163-b5d4-4575-a72c-e6ad1b3137df-kube-api-access-5cr4l\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.688901 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-hg688" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.688904 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-hg688" event={"ID":"9c985ff9-3545-4544-acc8-ca9eb2bfad6a","Type":"ContainerDied","Data":"485fdae7de5b84d520754700103a7b2d1a2699988fd5bac3363e984c2ccb2b0e"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.688952 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="485fdae7de5b84d520754700103a7b2d1a2699988fd5bac3363e984c2ccb2b0e" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.690835 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-ws5ht" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.690814 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-04a8-account-create-update-ws5ht" event={"ID":"37ba2d41-5de4-48cb-aad6-59df13307ed1","Type":"ContainerDied","Data":"85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.690913 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85ee0b6a029ba5152ff5abc51ef7d2c859d0da76a5bac6cb02c6728c11a62a9f" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.693142 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-cnnxs" event={"ID":"3f2b7163-b5d4-4575-a72c-e6ad1b3137df","Type":"ContainerDied","Data":"03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.693755 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03e0154228fb7cb7a65f349825e100c422b39b776d8d1f870768bf174633ebf0" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.693159 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-cnnxs" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.694515 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-v65tv" Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.694536 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-v65tv" event={"ID":"3db3f0b6-9cae-40fa-b54d-3ed06c568c9c","Type":"ContainerDied","Data":"46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553"} Mar 13 14:26:06 crc kubenswrapper[4907]: I0313 14:26:06.694568 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46e329b2292a23b4e7db5c2587420d8b2ff9fb0168254af0462a3f9acca4b553" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.024926 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.121323 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") pod \"7d930b98-c1af-4e47-a663-86afa9484856\" (UID: \"7d930b98-c1af-4e47-a663-86afa9484856\") " Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.138387 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44" (OuterVolumeSpecName: "kube-api-access-f5d44") pod "7d930b98-c1af-4e47-a663-86afa9484856" (UID: "7d930b98-c1af-4e47-a663-86afa9484856"). InnerVolumeSpecName "kube-api-access-f5d44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.225098 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5d44\" (UniqueName: \"kubernetes.io/projected/7d930b98-c1af-4e47-a663-86afa9484856-kube-api-access-f5d44\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.724017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556866-fwg48" event={"ID":"7d930b98-c1af-4e47-a663-86afa9484856","Type":"ContainerDied","Data":"c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9"} Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.724284 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c521a07f9fcca8ad2de643a0b0fa242b546a5a944f92129f2cbcb3aea70c71e9" Mar 13 14:26:07 crc kubenswrapper[4907]: I0313 14:26:07.724100 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556866-fwg48" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.081710 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.088377 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556860-ljdj7"] Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447178 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447511 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447525 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447539 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="init" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447547 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="init" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447563 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447571 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447610 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447622 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447635 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447642 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447663 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447670 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447682 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447689 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: E0313 14:26:08.447703 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d930b98-c1af-4e47-a663-86afa9484856" containerName="oc" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447710 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d930b98-c1af-4e47-a663-86afa9484856" containerName="oc" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447932 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447951 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="779b1958-f7ed-4c53-a4ff-a6c4a803bebf" containerName="dnsmasq-dns" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.447963 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12f5638-4cb2-4424-86d8-5c0c829d82d5" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448006 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" containerName="mariadb-account-create-update" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448018 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448033 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" containerName="mariadb-database-create" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448041 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d930b98-c1af-4e47-a663-86afa9484856" containerName="oc" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.448718 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.450797 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.456094 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.649522 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.649599 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.750862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.751599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.752621 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.767541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"root-account-create-update-r2rlh\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:08 crc kubenswrapper[4907]: I0313 14:26:08.768250 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:09 crc kubenswrapper[4907]: I0313 14:26:09.344322 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 13 14:26:09 crc kubenswrapper[4907]: I0313 14:26:09.804548 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="257ee294-461c-4344-9dcb-b44f1497f8bf" path="/var/lib/kubelet/pods/257ee294-461c-4344-9dcb-b44f1497f8bf/volumes" Mar 13 14:26:10 crc kubenswrapper[4907]: I0313 14:26:10.291710 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:10 crc kubenswrapper[4907]: E0313 14:26:10.291903 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:26:10 crc kubenswrapper[4907]: E0313 14:26:10.292184 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 14:26:10 crc kubenswrapper[4907]: E0313 14:26:10.292246 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:26:26.292227834 +0000 UTC m=+1285.192015523 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : configmap "swift-ring-files" not found Mar 13 14:26:10 crc kubenswrapper[4907]: I0313 14:26:10.770806 4907 generic.go:334] "Generic (PLEG): container finished" podID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerID="ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69" exitCode=0 Mar 13 14:26:10 crc kubenswrapper[4907]: I0313 14:26:10.770852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerDied","Data":"ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69"} Mar 13 14:26:12 crc kubenswrapper[4907]: I0313 14:26:12.472746 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" probeResult="failure" output=< Mar 13 14:26:12 crc kubenswrapper[4907]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 14:26:12 crc kubenswrapper[4907]: > Mar 13 14:26:12 crc kubenswrapper[4907]: I0313 14:26:12.481224 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:26:14 crc kubenswrapper[4907]: I0313 14:26:14.803361 4907 generic.go:334] "Generic (PLEG): container finished" podID="286c82e8-b74a-49d7-a355-ac074aace10e" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" exitCode=0 Mar 13 14:26:14 crc kubenswrapper[4907]: I0313 14:26:14.803470 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerDied","Data":"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7"} Mar 13 14:26:15 crc kubenswrapper[4907]: I0313 14:26:15.813676 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3c34454-315e-4821-ab25-b0f331a0d521" containerID="987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38" exitCode=0 Mar 13 14:26:15 crc kubenswrapper[4907]: I0313 14:26:15.813777 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerDied","Data":"987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38"} Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.467903 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" probeResult="failure" output=< Mar 13 14:26:17 crc kubenswrapper[4907]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 14:26:17 crc kubenswrapper[4907]: > Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.477975 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.682038 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.683015 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.685032 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.694851 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725618 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725684 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725708 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725806 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.725854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827711 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827727 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.827756 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828094 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828205 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.828993 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.830017 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:17 crc kubenswrapper[4907]: I0313 14:26:17.845315 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"ovn-controller-bw2xb-config-gdlfq\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:18 crc kubenswrapper[4907]: I0313 14:26:18.004090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:19 crc kubenswrapper[4907]: E0313 14:26:19.874211 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120" Mar 13 14:26:19 crc kubenswrapper[4907]: E0313 14:26:19.874636 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-65qwz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-4v4ps_openstack(02262347-eaec-462e-a3ef-58c670885d99): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:26:19 crc kubenswrapper[4907]: E0313 14:26:19.875851 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-4v4ps" podUID="02262347-eaec-462e-a3ef-58c670885d99" Mar 13 14:26:19 crc kubenswrapper[4907]: I0313 14:26:19.998155 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.062603 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.064291 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.064236 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.064372 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.065959 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066017 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066091 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066141 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") pod \"e2585497-6837-4d17-8d51-c3d7879fdb46\" (UID: \"e2585497-6837-4d17-8d51-c3d7879fdb46\") " Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.066796 4907 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e2585497-6837-4d17-8d51-c3d7879fdb46-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.068006 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.071135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86" (OuterVolumeSpecName: "kube-api-access-msr86") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "kube-api-access-msr86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.077729 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.097125 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.097721 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.107800 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts" (OuterVolumeSpecName: "scripts") pod "e2585497-6837-4d17-8d51-c3d7879fdb46" (UID: "e2585497-6837-4d17-8d51-c3d7879fdb46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.169835 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-msr86\" (UniqueName: \"kubernetes.io/projected/e2585497-6837-4d17-8d51-c3d7879fdb46-kube-api-access-msr86\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170107 4907 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170120 4907 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170128 4907 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170136 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e2585497-6837-4d17-8d51-c3d7879fdb46-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.170145 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2585497-6837-4d17-8d51-c3d7879fdb46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.332141 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.338656 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:20 crc kubenswrapper[4907]: W0313 14:26:20.342726 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e734847_721a_487c_ad5d_e947f9eb2a5b.slice/crio-f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b WatchSource:0}: Error finding container f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b: Status 404 returned error can't find the container with id f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b Mar 13 14:26:20 crc kubenswrapper[4907]: W0313 14:26:20.343405 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3affff5_a720_4d5e_8f17_e8ec23eff0c9.slice/crio-d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81 WatchSource:0}: Error finding container d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81: Status 404 returned error can't find the container with id d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81 Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.860350 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerStarted","Data":"c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.861009 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.863363 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerStarted","Data":"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.863620 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.883263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb-config-gdlfq" event={"ID":"8e734847-721a-487c-ad5d-e947f9eb2a5b","Type":"ContainerStarted","Data":"25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.883373 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb-config-gdlfq" event={"ID":"8e734847-721a-487c-ad5d-e947f9eb2a5b","Type":"ContainerStarted","Data":"f9b5f7112a5f19ae7b3025c0088305694e18947d40457ccaa65e79dd6ee2919b"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.890044 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=54.508031654 podStartE2EDuration="1m3.890018007s" podCreationTimestamp="2026-03-13 14:25:17 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.046998857 +0000 UTC m=+1229.946786546" lastFinishedPulling="2026-03-13 14:25:40.42898521 +0000 UTC m=+1239.328772899" observedRunningTime="2026-03-13 14:26:20.885361749 +0000 UTC m=+1279.785149458" watchObservedRunningTime="2026-03-13 14:26:20.890018007 +0000 UTC m=+1279.789805696" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.896543 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-9fxfh" event={"ID":"e2585497-6837-4d17-8d51-c3d7879fdb46","Type":"ContainerDied","Data":"c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.896593 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c44bcf911a2f35735f5bae70947997de7ad5c7bedca91be9607b903b91ea86c5" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.896618 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-9fxfh" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.898967 4907 generic.go:334] "Generic (PLEG): container finished" podID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerID="e6d88705fdc67e86fd0214f4c6bbf1c37e140065fdd712f700392f915f879795" exitCode=0 Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.899199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r2rlh" event={"ID":"b3affff5-a720-4d5e-8f17-e8ec23eff0c9","Type":"ContainerDied","Data":"e6d88705fdc67e86fd0214f4c6bbf1c37e140065fdd712f700392f915f879795"} Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.899240 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r2rlh" event={"ID":"b3affff5-a720-4d5e-8f17-e8ec23eff0c9","Type":"ContainerStarted","Data":"d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81"} Mar 13 14:26:20 crc kubenswrapper[4907]: E0313 14:26:20.905575 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:dae5e39780d5a15eed030c7009f8e5317139d447558ac83f038497be594be120\\\"\"" pod="openstack/glance-db-sync-4v4ps" podUID="02262347-eaec-462e-a3ef-58c670885d99" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.924042 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-bw2xb-config-gdlfq" podStartSLOduration=3.924020841 podStartE2EDuration="3.924020841s" podCreationTimestamp="2026-03-13 14:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:20.90285762 +0000 UTC m=+1279.802645309" watchObservedRunningTime="2026-03-13 14:26:20.924020841 +0000 UTC m=+1279.823808530" Mar 13 14:26:20 crc kubenswrapper[4907]: I0313 14:26:20.931657 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=56.717069131 podStartE2EDuration="1m4.93163531s" podCreationTimestamp="2026-03-13 14:25:16 +0000 UTC" firstStartedPulling="2026-03-13 14:25:31.607641697 +0000 UTC m=+1230.507429386" lastFinishedPulling="2026-03-13 14:25:39.822207876 +0000 UTC m=+1238.721995565" observedRunningTime="2026-03-13 14:26:20.922445088 +0000 UTC m=+1279.822232777" watchObservedRunningTime="2026-03-13 14:26:20.93163531 +0000 UTC m=+1279.831422999" Mar 13 14:26:21 crc kubenswrapper[4907]: I0313 14:26:21.907544 4907 generic.go:334] "Generic (PLEG): container finished" podID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerID="25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235" exitCode=0 Mar 13 14:26:21 crc kubenswrapper[4907]: I0313 14:26:21.907636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb-config-gdlfq" event={"ID":"8e734847-721a-487c-ad5d-e947f9eb2a5b","Type":"ContainerDied","Data":"25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235"} Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.192063 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.323599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") pod \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.323709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") pod \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\" (UID: \"b3affff5-a720-4d5e-8f17-e8ec23eff0c9\") " Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.324233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3affff5-a720-4d5e-8f17-e8ec23eff0c9" (UID: "b3affff5-a720-4d5e-8f17-e8ec23eff0c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.328738 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x" (OuterVolumeSpecName: "kube-api-access-pgg7x") pod "b3affff5-a720-4d5e-8f17-e8ec23eff0c9" (UID: "b3affff5-a720-4d5e-8f17-e8ec23eff0c9"). InnerVolumeSpecName "kube-api-access-pgg7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.425732 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgg7x\" (UniqueName: \"kubernetes.io/projected/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-kube-api-access-pgg7x\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.425776 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3affff5-a720-4d5e-8f17-e8ec23eff0c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.463832 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-bw2xb" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.919039 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r2rlh" Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.919046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r2rlh" event={"ID":"b3affff5-a720-4d5e-8f17-e8ec23eff0c9","Type":"ContainerDied","Data":"d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81"} Mar 13 14:26:22 crc kubenswrapper[4907]: I0313 14:26:22.919096 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d555d68605167e787613a0ff52207f2638755acf69400396ea72ba52afda2d81" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.214539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340416 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340512 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run" (OuterVolumeSpecName: "var-run") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340623 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340668 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340730 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") pod \"8e734847-721a-487c-ad5d-e947f9eb2a5b\" (UID: \"8e734847-721a-487c-ad5d-e947f9eb2a5b\") " Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340790 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.340909 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341084 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341426 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341459 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341473 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341487 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8e734847-721a-487c-ad5d-e947f9eb2a5b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.341610 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts" (OuterVolumeSpecName: "scripts") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.358171 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z" (OuterVolumeSpecName: "kube-api-access-9hp7z") pod "8e734847-721a-487c-ad5d-e947f9eb2a5b" (UID: "8e734847-721a-487c-ad5d-e947f9eb2a5b"). InnerVolumeSpecName "kube-api-access-9hp7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.434352 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.441020 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-bw2xb-config-gdlfq"] Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.442717 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hp7z\" (UniqueName: \"kubernetes.io/projected/8e734847-721a-487c-ad5d-e947f9eb2a5b-kube-api-access-9hp7z\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.442746 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8e734847-721a-487c-ad5d-e947f9eb2a5b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.791927 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" path="/var/lib/kubelet/pods/8e734847-721a-487c-ad5d-e947f9eb2a5b/volumes" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.927634 4907 scope.go:117] "RemoveContainer" containerID="25b2791fa541284bf13b0da85be9466979de89ae098c8a967456ebbd1cd18235" Mar 13 14:26:23 crc kubenswrapper[4907]: I0313 14:26:23.927774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb-config-gdlfq" Mar 13 14:26:24 crc kubenswrapper[4907]: I0313 14:26:24.825962 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:24 crc kubenswrapper[4907]: I0313 14:26:24.833671 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-r2rlh"] Mar 13 14:26:25 crc kubenswrapper[4907]: I0313 14:26:25.791507 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" path="/var/lib/kubelet/pods/b3affff5-a720-4d5e-8f17-e8ec23eff0c9/volumes" Mar 13 14:26:26 crc kubenswrapper[4907]: I0313 14:26:26.386452 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:26 crc kubenswrapper[4907]: I0313 14:26:26.393420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"swift-storage-0\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " pod="openstack/swift-storage-0" Mar 13 14:26:26 crc kubenswrapper[4907]: I0313 14:26:26.461809 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:26:27 crc kubenswrapper[4907]: I0313 14:26:27.288909 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:26:27 crc kubenswrapper[4907]: I0313 14:26:27.290222 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:26:27 crc kubenswrapper[4907]: I0313 14:26:27.966606 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"8ab726bec0bb062d6ac5d70e873cfd78800b9921e7163b9cb0ca74467cc3c7f4"} Mar 13 14:26:28 crc kubenswrapper[4907]: I0313 14:26:28.976010 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091"} Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.845819 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:26:29 crc kubenswrapper[4907]: E0313 14:26:29.846413 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerName="ovn-config" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848246 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerName="ovn-config" Mar 13 14:26:29 crc kubenswrapper[4907]: E0313 14:26:29.848297 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerName="mariadb-account-create-update" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848305 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerName="mariadb-account-create-update" Mar 13 14:26:29 crc kubenswrapper[4907]: E0313 14:26:29.848318 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerName="swift-ring-rebalance" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848325 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerName="swift-ring-rebalance" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848495 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3affff5-a720-4d5e-8f17-e8ec23eff0c9" containerName="mariadb-account-create-update" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848516 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" containerName="swift-ring-rebalance" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.848529 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e734847-721a-487c-ad5d-e947f9eb2a5b" containerName="ovn-config" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.849083 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.851597 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.857019 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.940379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.940478 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.989092 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410"} Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.989140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e"} Mar 13 14:26:29 crc kubenswrapper[4907]: I0313 14:26:29.989153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566"} Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.042765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.042832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.044589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.064831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"root-account-create-update-r6vk2\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.168017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.431260 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.997095 4907 generic.go:334] "Generic (PLEG): container finished" podID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerID="3e1e04ee01cb4c593a590da900f137153aed8f7c58309d1a24f19dd94c9af998" exitCode=0 Mar 13 14:26:30 crc kubenswrapper[4907]: I0313 14:26:30.997281 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r6vk2" event={"ID":"d1b4fd44-4a11-4521-bfd0-16c5436368fa","Type":"ContainerDied","Data":"3e1e04ee01cb4c593a590da900f137153aed8f7c58309d1a24f19dd94c9af998"} Mar 13 14:26:31 crc kubenswrapper[4907]: I0313 14:26:30.997724 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r6vk2" event={"ID":"d1b4fd44-4a11-4521-bfd0-16c5436368fa","Type":"ContainerStarted","Data":"2471a799ca22d8e83409150033938766a0e509f65d736ee99f5f36e665621536"} Mar 13 14:26:31 crc kubenswrapper[4907]: I0313 14:26:31.001769 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80"} Mar 13 14:26:31 crc kubenswrapper[4907]: I0313 14:26:31.001818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a"} Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.012398 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f"} Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.012795 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808"} Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.303672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.383535 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") pod \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.383666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") pod \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\" (UID: \"d1b4fd44-4a11-4521-bfd0-16c5436368fa\") " Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.384250 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1b4fd44-4a11-4521-bfd0-16c5436368fa" (UID: "d1b4fd44-4a11-4521-bfd0-16c5436368fa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.388974 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5" (OuterVolumeSpecName: "kube-api-access-grms5") pod "d1b4fd44-4a11-4521-bfd0-16c5436368fa" (UID: "d1b4fd44-4a11-4521-bfd0-16c5436368fa"). InnerVolumeSpecName "kube-api-access-grms5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.485625 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grms5\" (UniqueName: \"kubernetes.io/projected/d1b4fd44-4a11-4521-bfd0-16c5436368fa-kube-api-access-grms5\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:32 crc kubenswrapper[4907]: I0313 14:26:32.485946 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b4fd44-4a11-4521-bfd0-16c5436368fa-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.040111 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-r6vk2" event={"ID":"d1b4fd44-4a11-4521-bfd0-16c5436368fa","Type":"ContainerDied","Data":"2471a799ca22d8e83409150033938766a0e509f65d736ee99f5f36e665621536"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.040159 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2471a799ca22d8e83409150033938766a0e509f65d736ee99f5f36e665621536" Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.040213 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-r6vk2" Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061270 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061315 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061342 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94"} Mar 13 14:26:33 crc kubenswrapper[4907]: I0313 14:26:33.061354 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7"} Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.076850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea"} Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.077245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerStarted","Data":"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411"} Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.113198 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.27268048 podStartE2EDuration="41.113178399s" podCreationTimestamp="2026-03-13 14:25:53 +0000 UTC" firstStartedPulling="2026-03-13 14:26:27.288573833 +0000 UTC m=+1286.188361532" lastFinishedPulling="2026-03-13 14:26:32.129071762 +0000 UTC m=+1291.028859451" observedRunningTime="2026-03-13 14:26:34.110913706 +0000 UTC m=+1293.010701425" watchObservedRunningTime="2026-03-13 14:26:34.113178399 +0000 UTC m=+1293.012966098" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.396971 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:26:34 crc kubenswrapper[4907]: E0313 14:26:34.397584 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerName="mariadb-account-create-update" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.397652 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerName="mariadb-account-create-update" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.397851 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" containerName="mariadb-account-create-update" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.398746 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.400931 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.413128 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520241 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520310 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520377 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520395 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.520412 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622062 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622141 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622264 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.622290 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623217 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.623915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.624038 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.641135 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"dnsmasq-dns-67754df655-k2v49\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:34 crc kubenswrapper[4907]: I0313 14:26:34.718253 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:35 crc kubenswrapper[4907]: I0313 14:26:35.157264 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.098490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerStarted","Data":"2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4"} Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.100551 4907 generic.go:334] "Generic (PLEG): container finished" podID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerID="fc5cca7a53563949cc0a2a52bb6bc5ab553013ec2df261991d02d7a2bd1a082d" exitCode=0 Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.100608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerDied","Data":"fc5cca7a53563949cc0a2a52bb6bc5ab553013ec2df261991d02d7a2bd1a082d"} Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.100647 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerStarted","Data":"1dc89b2197970bff5fe57426501341e047b5f1d0b4ba93884702cc3864ea9041"} Mar 13 14:26:36 crc kubenswrapper[4907]: I0313 14:26:36.145838 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4v4ps" podStartSLOduration=3.156574488 podStartE2EDuration="35.145814487s" podCreationTimestamp="2026-03-13 14:26:01 +0000 UTC" firstStartedPulling="2026-03-13 14:26:03.2710231 +0000 UTC m=+1262.170810789" lastFinishedPulling="2026-03-13 14:26:35.260263099 +0000 UTC m=+1294.160050788" observedRunningTime="2026-03-13 14:26:36.121278114 +0000 UTC m=+1295.021065803" watchObservedRunningTime="2026-03-13 14:26:36.145814487 +0000 UTC m=+1295.045602186" Mar 13 14:26:37 crc kubenswrapper[4907]: I0313 14:26:37.109992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerStarted","Data":"fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213"} Mar 13 14:26:37 crc kubenswrapper[4907]: I0313 14:26:37.110369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:37 crc kubenswrapper[4907]: I0313 14:26:37.140686 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67754df655-k2v49" podStartSLOduration=3.140657178 podStartE2EDuration="3.140657178s" podCreationTimestamp="2026-03-13 14:26:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:37.132839033 +0000 UTC m=+1296.032626802" watchObservedRunningTime="2026-03-13 14:26:37.140657178 +0000 UTC m=+1296.040444897" Mar 13 14:26:38 crc kubenswrapper[4907]: I0313 14:26:38.278097 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:26:38 crc kubenswrapper[4907]: I0313 14:26:38.698068 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.120125 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.121328 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.143458 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.215554 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.216828 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.217556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.217616 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.219470 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.230057 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.231497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.241226 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.250441 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319498 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319578 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319607 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319677 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.319714 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.320338 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.348628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"cinder-db-create-vbcql\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.411289 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.412544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422603 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422746 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.422831 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.424819 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.439419 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.444429 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.445638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.484669 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"cinder-47e2-account-create-update-r4ccm\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.485907 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.487381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.488026 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"barbican-db-create-htpvb\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.489973 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.490237 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.490596 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.490875 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.497072 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.498336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.500092 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.504676 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.515401 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.524389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.524450 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.532990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.545772 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.626299 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.626764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.626838 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627083 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627162 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.627298 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.628421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.641041 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.642544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.662778 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.684212 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.703465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"neutron-db-create-x9b2b\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728454 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728537 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728561 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728622 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728641 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.728675 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.729418 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.735363 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.735984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.739219 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.754523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"keystone-db-sync-kcjn8\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.762532 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"barbican-b501-account-create-update-dzlqr\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.830143 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.830217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.832312 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.853058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"neutron-b69e-account-create-update-ptpcx\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.964952 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:40 crc kubenswrapper[4907]: I0313 14:26:40.982351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.003092 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.051118 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.181390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.184225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vbcql" event={"ID":"e1068e89-10df-4d4c-a213-d9e9643fab38","Type":"ContainerStarted","Data":"800ea320e37f9ac3fbdc6c79cbb80396598682384d93ce308ede98d248a09cc4"} Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.211675 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:26:41 crc kubenswrapper[4907]: W0313 14:26:41.242715 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda89525da_db62_49c2_b548_bab4c974dfdd.slice/crio-265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d WatchSource:0}: Error finding container 265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d: Status 404 returned error can't find the container with id 265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.269592 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:26:41 crc kubenswrapper[4907]: W0313 14:26:41.284063 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bab6ed6_5335_40cc_bdff_3d7461876fb8.slice/crio-ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3 WatchSource:0}: Error finding container ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3: Status 404 returned error can't find the container with id ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3 Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.517999 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:26:41 crc kubenswrapper[4907]: W0313 14:26:41.527091 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0983a1bd_a83c_4658_a405_09b8ab0d0002.slice/crio-807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567 WatchSource:0}: Error finding container 807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567: Status 404 returned error can't find the container with id 807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567 Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.594557 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:26:41 crc kubenswrapper[4907]: I0313 14:26:41.629836 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.198530 4907 generic.go:334] "Generic (PLEG): container finished" podID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerID="8f4cb6476a276775b25f72fa46f11e3f550f531236e5cf19234f8b274637da17" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.198871 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-r4ccm" event={"ID":"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83","Type":"ContainerDied","Data":"8f4cb6476a276775b25f72fa46f11e3f550f531236e5cf19234f8b274637da17"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.198914 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-r4ccm" event={"ID":"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83","Type":"ContainerStarted","Data":"31fc0aff0eaffc735f077ff239eae73496c2aa5d04daeeff77d6a777c8e6f3f3"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.200453 4907 generic.go:334] "Generic (PLEG): container finished" podID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerID="cdd8592044b8b42302c000b60e593bcd7aa7352aa0bfd95a951e13ddb4d1cef1" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.200490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b501-account-create-update-dzlqr" event={"ID":"98748d4d-f934-49ed-91d9-7a9e467b2753","Type":"ContainerDied","Data":"cdd8592044b8b42302c000b60e593bcd7aa7352aa0bfd95a951e13ddb4d1cef1"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.200505 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b501-account-create-update-dzlqr" event={"ID":"98748d4d-f934-49ed-91d9-7a9e467b2753","Type":"ContainerStarted","Data":"5a33722749134dbbc6eb786e1dd47e935db2b7e8fc7b1822f79e1da556bcd6c2"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.203361 4907 generic.go:334] "Generic (PLEG): container finished" podID="a89525da-db62-49c2-b548-bab4c974dfdd" containerID="766d1e6536dea7ad49da9b30346790d8d4eb902b9e7e7b3c7bf460cfd415e154" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.203417 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htpvb" event={"ID":"a89525da-db62-49c2-b548-bab4c974dfdd","Type":"ContainerDied","Data":"766d1e6536dea7ad49da9b30346790d8d4eb902b9e7e7b3c7bf460cfd415e154"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.203440 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htpvb" event={"ID":"a89525da-db62-49c2-b548-bab4c974dfdd","Type":"ContainerStarted","Data":"265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.206535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerStarted","Data":"35b6cb9082e770b5ec87125f38e9a5f5d08513cff5fef490907772e3d539df90"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.208310 4907 generic.go:334] "Generic (PLEG): container finished" podID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerID="0a3f47d3069fe894c10b414d1b636502b45151a11f2acdaa45abe3e56ee5a7a1" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.208360 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vbcql" event={"ID":"e1068e89-10df-4d4c-a213-d9e9643fab38","Type":"ContainerDied","Data":"0a3f47d3069fe894c10b414d1b636502b45151a11f2acdaa45abe3e56ee5a7a1"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.210605 4907 generic.go:334] "Generic (PLEG): container finished" podID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerID="71b8e65c6bc18f48a724ed12b9dcf478104abf3f0beba637548e1962f43f12bb" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.210686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9b2b" event={"ID":"2bab6ed6-5335-40cc-bdff-3d7461876fb8","Type":"ContainerDied","Data":"71b8e65c6bc18f48a724ed12b9dcf478104abf3f0beba637548e1962f43f12bb"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.210714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9b2b" event={"ID":"2bab6ed6-5335-40cc-bdff-3d7461876fb8","Type":"ContainerStarted","Data":"ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.218261 4907 generic.go:334] "Generic (PLEG): container finished" podID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerID="e3a0771ba1e8115d911c740ed70ee5fca37f11ff18ec41bfab54c46258690b83" exitCode=0 Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.218317 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-ptpcx" event={"ID":"0983a1bd-a83c-4658-a405-09b8ab0d0002","Type":"ContainerDied","Data":"e3a0771ba1e8115d911c740ed70ee5fca37f11ff18ec41bfab54c46258690b83"} Mar 13 14:26:42 crc kubenswrapper[4907]: I0313 14:26:42.218343 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-ptpcx" event={"ID":"0983a1bd-a83c-4658-a405-09b8ab0d0002","Type":"ContainerStarted","Data":"807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567"} Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.235941 4907 generic.go:334] "Generic (PLEG): container finished" podID="02262347-eaec-462e-a3ef-58c670885d99" containerID="2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4" exitCode=0 Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.236052 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerDied","Data":"2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4"} Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.720064 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.808530 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:26:44 crc kubenswrapper[4907]: I0313 14:26:44.808753 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" containerID="cri-o://ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96" gracePeriod=10 Mar 13 14:26:45 crc kubenswrapper[4907]: I0313 14:26:45.256559 4907 generic.go:334] "Generic (PLEG): container finished" podID="89cd945f-311f-47d1-982a-641f062e4f57" containerID="ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96" exitCode=0 Mar 13 14:26:45 crc kubenswrapper[4907]: I0313 14:26:45.256644 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerDied","Data":"ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96"} Mar 13 14:26:46 crc kubenswrapper[4907]: I0313 14:26:46.996723 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.004938 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.020855 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.030322 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.041981 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.063183 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") pod \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088176 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") pod \"98748d4d-f934-49ed-91d9-7a9e467b2753\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088246 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") pod \"98748d4d-f934-49ed-91d9-7a9e467b2753\" (UID: \"98748d4d-f934-49ed-91d9-7a9e467b2753\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.088308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") pod \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\" (UID: \"2bab6ed6-5335-40cc-bdff-3d7461876fb8\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.089640 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98748d4d-f934-49ed-91d9-7a9e467b2753" (UID: "98748d4d-f934-49ed-91d9-7a9e467b2753"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.090695 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2bab6ed6-5335-40cc-bdff-3d7461876fb8" (UID: "2bab6ed6-5335-40cc-bdff-3d7461876fb8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.102767 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d" (OuterVolumeSpecName: "kube-api-access-lwp6d") pod "2bab6ed6-5335-40cc-bdff-3d7461876fb8" (UID: "2bab6ed6-5335-40cc-bdff-3d7461876fb8"). InnerVolumeSpecName "kube-api-access-lwp6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.104507 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb" (OuterVolumeSpecName: "kube-api-access-ddmjb") pod "98748d4d-f934-49ed-91d9-7a9e467b2753" (UID: "98748d4d-f934-49ed-91d9-7a9e467b2753"). InnerVolumeSpecName "kube-api-access-ddmjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191075 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") pod \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") pod \"a89525da-db62-49c2-b548-bab4c974dfdd\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191206 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") pod \"0983a1bd-a83c-4658-a405-09b8ab0d0002\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191244 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") pod \"0983a1bd-a83c-4658-a405-09b8ab0d0002\" (UID: \"0983a1bd-a83c-4658-a405-09b8ab0d0002\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191289 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") pod \"e1068e89-10df-4d4c-a213-d9e9643fab38\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191328 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") pod \"e1068e89-10df-4d4c-a213-d9e9643fab38\" (UID: \"e1068e89-10df-4d4c-a213-d9e9643fab38\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191353 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") pod \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\" (UID: \"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.191384 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") pod \"a89525da-db62-49c2-b548-bab4c974dfdd\" (UID: \"a89525da-db62-49c2-b548-bab4c974dfdd\") " Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194018 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwp6d\" (UniqueName: \"kubernetes.io/projected/2bab6ed6-5335-40cc-bdff-3d7461876fb8-kube-api-access-lwp6d\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194045 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98748d4d-f934-49ed-91d9-7a9e467b2753-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194057 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddmjb\" (UniqueName: \"kubernetes.io/projected/98748d4d-f934-49ed-91d9-7a9e467b2753-kube-api-access-ddmjb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194069 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bab6ed6-5335-40cc-bdff-3d7461876fb8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194249 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" (UID: "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194471 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a89525da-db62-49c2-b548-bab4c974dfdd" (UID: "a89525da-db62-49c2-b548-bab4c974dfdd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194596 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1068e89-10df-4d4c-a213-d9e9643fab38" (UID: "e1068e89-10df-4d4c-a213-d9e9643fab38"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.194971 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0983a1bd-a83c-4658-a405-09b8ab0d0002" (UID: "0983a1bd-a83c-4658-a405-09b8ab0d0002"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.199709 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp" (OuterVolumeSpecName: "kube-api-access-kgvgp") pod "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" (UID: "3fe8116b-9e68-41e2-aad2-cdd7d5e51c83"). InnerVolumeSpecName "kube-api-access-kgvgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.204428 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5" (OuterVolumeSpecName: "kube-api-access-d2gv5") pod "0983a1bd-a83c-4658-a405-09b8ab0d0002" (UID: "0983a1bd-a83c-4658-a405-09b8ab0d0002"). InnerVolumeSpecName "kube-api-access-d2gv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.210805 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks" (OuterVolumeSpecName: "kube-api-access-qv2ks") pod "a89525da-db62-49c2-b548-bab4c974dfdd" (UID: "a89525da-db62-49c2-b548-bab4c974dfdd"). InnerVolumeSpecName "kube-api-access-qv2ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.232215 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd" (OuterVolumeSpecName: "kube-api-access-chjsd") pod "e1068e89-10df-4d4c-a213-d9e9643fab38" (UID: "e1068e89-10df-4d4c-a213-d9e9643fab38"). InnerVolumeSpecName "kube-api-access-chjsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.280762 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-r4ccm" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.280994 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-r4ccm" event={"ID":"3fe8116b-9e68-41e2-aad2-cdd7d5e51c83","Type":"ContainerDied","Data":"31fc0aff0eaffc735f077ff239eae73496c2aa5d04daeeff77d6a777c8e6f3f3"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.281030 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31fc0aff0eaffc735f077ff239eae73496c2aa5d04daeeff77d6a777c8e6f3f3" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.283943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b501-account-create-update-dzlqr" event={"ID":"98748d4d-f934-49ed-91d9-7a9e467b2753","Type":"ContainerDied","Data":"5a33722749134dbbc6eb786e1dd47e935db2b7e8fc7b1822f79e1da556bcd6c2"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.283968 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a33722749134dbbc6eb786e1dd47e935db2b7e8fc7b1822f79e1da556bcd6c2" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.284106 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b501-account-create-update-dzlqr" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.285689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-htpvb" event={"ID":"a89525da-db62-49c2-b548-bab4c974dfdd","Type":"ContainerDied","Data":"265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.285711 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="265dfd02ff62990e5c06a12bdcd62f177d8601a264d83ff1c46d7bdee007b20d" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.285783 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-htpvb" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.294364 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vbcql" event={"ID":"e1068e89-10df-4d4c-a213-d9e9643fab38","Type":"ContainerDied","Data":"800ea320e37f9ac3fbdc6c79cbb80396598682384d93ce308ede98d248a09cc4"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.294453 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="800ea320e37f9ac3fbdc6c79cbb80396598682384d93ce308ede98d248a09cc4" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.294390 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vbcql" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295750 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2gv5\" (UniqueName: \"kubernetes.io/projected/0983a1bd-a83c-4658-a405-09b8ab0d0002-kube-api-access-d2gv5\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295773 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0983a1bd-a83c-4658-a405-09b8ab0d0002-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295787 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1068e89-10df-4d4c-a213-d9e9643fab38-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295800 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chjsd\" (UniqueName: \"kubernetes.io/projected/e1068e89-10df-4d4c-a213-d9e9643fab38-kube-api-access-chjsd\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295813 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgvgp\" (UniqueName: \"kubernetes.io/projected/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-kube-api-access-kgvgp\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295829 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qv2ks\" (UniqueName: \"kubernetes.io/projected/a89525da-db62-49c2-b548-bab4c974dfdd-kube-api-access-qv2ks\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295847 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a89525da-db62-49c2-b548-bab4c974dfdd-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.295859 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.298835 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9b2b" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.298836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9b2b" event={"ID":"2bab6ed6-5335-40cc-bdff-3d7461876fb8","Type":"ContainerDied","Data":"ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.299827 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce59465e7c0c37d5e78037bb1b277bb43931b1fc4a590e4e9d4e42dd7fdc27d3" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.317221 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-ptpcx" event={"ID":"0983a1bd-a83c-4658-a405-09b8ab0d0002","Type":"ContainerDied","Data":"807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567"} Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.317270 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="807b6a972511f96035dc8e7a5888a9cb76102f79128af800532f9871f36ec567" Mar 13 14:26:47 crc kubenswrapper[4907]: I0313 14:26:47.317295 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-ptpcx" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.121261 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.139669 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212231 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212343 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212385 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.212448 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") pod \"89cd945f-311f-47d1-982a-641f062e4f57\" (UID: \"89cd945f-311f-47d1-982a-641f062e4f57\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.229073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5" (OuterVolumeSpecName: "kube-api-access-lnkf5") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "kube-api-access-lnkf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.256482 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.257177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config" (OuterVolumeSpecName: "config") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.265075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.278276 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "89cd945f-311f-47d1-982a-641f062e4f57" (UID: "89cd945f-311f-47d1-982a-641f062e4f57"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.325710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.325857 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.325920 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") pod \"02262347-eaec-462e-a3ef-58c670885d99\" (UID: \"02262347-eaec-462e-a3ef-58c670885d99\") " Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326611 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326657 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326672 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326705 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89cd945f-311f-47d1-982a-641f062e4f57-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.326739 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnkf5\" (UniqueName: \"kubernetes.io/projected/89cd945f-311f-47d1-982a-641f062e4f57-kube-api-access-lnkf5\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.327636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" event={"ID":"89cd945f-311f-47d1-982a-641f062e4f57","Type":"ContainerDied","Data":"b45d10603df2d7ee66cbd50a38052eed85880ecd3d8f94e817d4236b2c83c476"} Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.327681 4907 scope.go:117] "RemoveContainer" containerID="ad9755b55eba4bb5e259a681410af3ffba2c66af2583b34cd43ac29178757a96" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.327816 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b9fd7d84c-wkdxf" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.329220 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.331602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz" (OuterVolumeSpecName: "kube-api-access-65qwz") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "kube-api-access-65qwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.334259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4v4ps" event={"ID":"02262347-eaec-462e-a3ef-58c670885d99","Type":"ContainerDied","Data":"e0469ec6298d86f707976b7020c89f6d29a4be70c9a77f315ceb22e64cbb4824"} Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.334336 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0469ec6298d86f707976b7020c89f6d29a4be70c9a77f315ceb22e64cbb4824" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.334350 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4v4ps" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.350408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.365619 4907 scope.go:117] "RemoveContainer" containerID="b1318d08ab870fbe30d1438d26843f78b5e12e9bf7ba05a68e44fb095db6996e" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.367080 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.369162 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data" (OuterVolumeSpecName: "config-data") pod "02262347-eaec-462e-a3ef-58c670885d99" (UID: "02262347-eaec-462e-a3ef-58c670885d99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.376563 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b9fd7d84c-wkdxf"] Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428371 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65qwz\" (UniqueName: \"kubernetes.io/projected/02262347-eaec-462e-a3ef-58c670885d99-kube-api-access-65qwz\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428632 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428715 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:48 crc kubenswrapper[4907]: I0313 14:26:48.428799 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02262347-eaec-462e-a3ef-58c670885d99-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.346554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerStarted","Data":"fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933"} Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.366343 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-kcjn8" podStartSLOduration=2.737317143 podStartE2EDuration="9.366322105s" podCreationTimestamp="2026-03-13 14:26:40 +0000 UTC" firstStartedPulling="2026-03-13 14:26:41.671021732 +0000 UTC m=+1300.570809421" lastFinishedPulling="2026-03-13 14:26:48.300026694 +0000 UTC m=+1307.199814383" observedRunningTime="2026-03-13 14:26:49.363072176 +0000 UTC m=+1308.262859885" watchObservedRunningTime="2026-03-13 14:26:49.366322105 +0000 UTC m=+1308.266109814" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676090 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676672 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="init" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676689 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="init" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676697 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676703 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676717 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676725 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676735 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676741 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676753 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676759 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676770 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676776 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676787 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676793 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676805 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02262347-eaec-462e-a3ef-58c670885d99" containerName="glance-db-sync" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676811 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="02262347-eaec-462e-a3ef-58c670885d99" containerName="glance-db-sync" Mar 13 14:26:49 crc kubenswrapper[4907]: E0313 14:26:49.676831 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.676837 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677025 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="89cd945f-311f-47d1-982a-641f062e4f57" containerName="dnsmasq-dns" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677036 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677048 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677059 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677073 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" containerName="mariadb-account-create-update" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677083 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677093 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="02262347-eaec-462e-a3ef-58c670885d99" containerName="glance-db-sync" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677100 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" containerName="mariadb-database-create" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.677908 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.695106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.792908 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89cd945f-311f-47d1-982a-641f062e4f57" path="/var/lib/kubelet/pods/89cd945f-311f-47d1-982a-641f062e4f57/volumes" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853115 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853611 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.853936 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.955769 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956223 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956482 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.956843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.957429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.957456 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.957794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:49 crc kubenswrapper[4907]: I0313 14:26:49.958114 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:50 crc kubenswrapper[4907]: I0313 14:26:50.011782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"dnsmasq-dns-6f88567fd9-f24rf\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:50 crc kubenswrapper[4907]: I0313 14:26:50.299492 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:50 crc kubenswrapper[4907]: I0313 14:26:50.781404 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:51 crc kubenswrapper[4907]: I0313 14:26:51.393028 4907 generic.go:334] "Generic (PLEG): container finished" podID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerID="b20445b78457705b4b60b2201ff58c281da6086f771d56be4ceb3ab98dc8b1f5" exitCode=0 Mar 13 14:26:51 crc kubenswrapper[4907]: I0313 14:26:51.393124 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerDied","Data":"b20445b78457705b4b60b2201ff58c281da6086f771d56be4ceb3ab98dc8b1f5"} Mar 13 14:26:51 crc kubenswrapper[4907]: I0313 14:26:51.393340 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerStarted","Data":"6f3c01688bba973a49cd38ead6f32071203a13f1c437e87590a2062f43c42d3e"} Mar 13 14:26:52 crc kubenswrapper[4907]: I0313 14:26:52.402456 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerStarted","Data":"3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012"} Mar 13 14:26:52 crc kubenswrapper[4907]: I0313 14:26:52.402854 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:52 crc kubenswrapper[4907]: I0313 14:26:52.426408 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" podStartSLOduration=3.42638791 podStartE2EDuration="3.42638791s" podCreationTimestamp="2026-03-13 14:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:52.424469808 +0000 UTC m=+1311.324257497" watchObservedRunningTime="2026-03-13 14:26:52.42638791 +0000 UTC m=+1311.326175599" Mar 13 14:26:54 crc kubenswrapper[4907]: I0313 14:26:54.431821 4907 generic.go:334] "Generic (PLEG): container finished" podID="480d23fe-2980-46b3-9e86-e74f2d051357" containerID="fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933" exitCode=0 Mar 13 14:26:54 crc kubenswrapper[4907]: I0313 14:26:54.431932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerDied","Data":"fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933"} Mar 13 14:26:55 crc kubenswrapper[4907]: I0313 14:26:55.887999 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.018142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") pod \"480d23fe-2980-46b3-9e86-e74f2d051357\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.018326 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") pod \"480d23fe-2980-46b3-9e86-e74f2d051357\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.018517 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") pod \"480d23fe-2980-46b3-9e86-e74f2d051357\" (UID: \"480d23fe-2980-46b3-9e86-e74f2d051357\") " Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.024212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx" (OuterVolumeSpecName: "kube-api-access-g78kx") pod "480d23fe-2980-46b3-9e86-e74f2d051357" (UID: "480d23fe-2980-46b3-9e86-e74f2d051357"). InnerVolumeSpecName "kube-api-access-g78kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.041336 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "480d23fe-2980-46b3-9e86-e74f2d051357" (UID: "480d23fe-2980-46b3-9e86-e74f2d051357"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.062107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data" (OuterVolumeSpecName: "config-data") pod "480d23fe-2980-46b3-9e86-e74f2d051357" (UID: "480d23fe-2980-46b3-9e86-e74f2d051357"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.120632 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g78kx\" (UniqueName: \"kubernetes.io/projected/480d23fe-2980-46b3-9e86-e74f2d051357-kube-api-access-g78kx\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.120666 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.120677 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d23fe-2980-46b3-9e86-e74f2d051357-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.456331 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kcjn8" event={"ID":"480d23fe-2980-46b3-9e86-e74f2d051357","Type":"ContainerDied","Data":"35b6cb9082e770b5ec87125f38e9a5f5d08513cff5fef490907772e3d539df90"} Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.456381 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35b6cb9082e770b5ec87125f38e9a5f5d08513cff5fef490907772e3d539df90" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.456473 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kcjn8" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.763929 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:26:56 crc kubenswrapper[4907]: E0313 14:26:56.764358 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" containerName="keystone-db-sync" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.764373 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" containerName="keystone-db-sync" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.764589 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" containerName="keystone-db-sync" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.765275 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.770938 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.771304 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.771360 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.771453 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.773313 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.799378 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.799634 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" containerID="cri-o://3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012" gracePeriod=10 Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.801995 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.816515 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.864577 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.865925 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.895233 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937736 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937960 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.937992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:56 crc kubenswrapper[4907]: I0313 14:26:56.938077 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.050861 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.050956 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.050984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051015 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051083 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051108 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051135 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051172 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051241 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.051261 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.059661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.062234 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.062742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.070404 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.080679 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.085051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.087287 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"keystone-bootstrap-4tskm\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.087761 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.114311 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.114627 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.152785 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153864 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153923 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153946 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.153996 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.154668 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.154837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.155212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.180020 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.210907 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"dnsmasq-dns-5cb4dcfdd7-k6pkx\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.212495 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.213483 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.220372 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.222168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.228860 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.230091 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.235686 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.235987 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.236174 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-skqjb" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.236391 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d7phq" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.236440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.241372 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.242840 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.248127 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.248357 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tvcdf" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.248505 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.251570 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256812 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256852 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256919 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256950 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.256988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.257004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.304677 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.309346 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.314453 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.323985 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.324080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.324347 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pg28b" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.332460 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358070 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358094 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358119 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358144 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358169 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358219 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358258 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358293 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358414 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358447 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.358474 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.359372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.366172 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.369345 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.371813 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.372794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.374415 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.383777 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.408520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"ceilometer-0\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.434248 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.435600 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.458567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459286 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459309 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459334 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459380 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459437 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459474 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459496 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459573 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459689 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.459763 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.471224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.473816 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.477749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.492349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.507687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.512513 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.512922 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.515376 4907 generic.go:334] "Generic (PLEG): container finished" podID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerID="3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012" exitCode=0 Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.515898 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.515893 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerDied","Data":"3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012"} Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.516085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.523532 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"barbican-db-sync-rz6qh\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.523581 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"neutron-db-sync-jt88g\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.539292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"cinder-db-sync-7q999\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560851 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560945 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560969 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.560997 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561203 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561218 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.561243 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.567513 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.567670 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.569255 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.574927 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.583529 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"placement-db-sync-jjqr9\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.594222 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.641589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.674864 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.674953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675048 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675085 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675130 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.675199 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.676162 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.676224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.676830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.677015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.677772 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.683060 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.694786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"dnsmasq-dns-759cc7f497-7nv78\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.710645 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.721140 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.754045 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778776 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778828 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778932 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.778976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.779047 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.779068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") pod \"7aba6c70-a80a-4bce-9216-887cb7a2061c\" (UID: \"7aba6c70-a80a-4bce-9216-887cb7a2061c\") " Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.787112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z" (OuterVolumeSpecName: "kube-api-access-2dk2z") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "kube-api-access-2dk2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.875819 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config" (OuterVolumeSpecName: "config") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.881632 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.881674 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dk2z\" (UniqueName: \"kubernetes.io/projected/7aba6c70-a80a-4bce-9216-887cb7a2061c-kube-api-access-2dk2z\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.886972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.892778 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.900815 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.952963 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: E0313 14:26:57.953408 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.953424 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" Mar 13 14:26:57 crc kubenswrapper[4907]: E0313 14:26:57.953443 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="init" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.953451 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="init" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.953654 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" containerName="dnsmasq-dns" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.954682 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.960511 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.960856 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.961089 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.961838 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-fc9j2" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.972042 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.976859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7aba6c70-a80a-4bce-9216-887cb7a2061c" (UID: "7aba6c70-a80a-4bce-9216-887cb7a2061c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.978136 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983386 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983424 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983439 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:57 crc kubenswrapper[4907]: I0313 14:26:57.983451 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7aba6c70-a80a-4bce-9216-887cb7a2061c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.032990 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.036095 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.037717 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.041129 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.054220 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.076802 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085627 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085926 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085953 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.085985 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.086035 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.086069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.086092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187505 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187561 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187594 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187611 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.187650 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191691 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191800 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191820 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.191839 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.192013 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.192065 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.192087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.194232 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.212133 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.213074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.218144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.256198 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.278624 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.283570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.286937 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300440 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300556 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300624 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300720 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.300790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.302141 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.303106 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.308625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.308913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.314776 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.325210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.325856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.327851 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.328111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.342968 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.400189 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.531703 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerStarted","Data":"8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.532093 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerStarted","Data":"a2858c30fa2ae4911b2a7afe4d6fb52407d5d3328c466096a4d4cc1761524119"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.535681 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"cf9fd571ce09eea426c8836be80b3c4ef776ec41dabdcd7886212b3709f230ce"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.552327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" event={"ID":"7aba6c70-a80a-4bce-9216-887cb7a2061c","Type":"ContainerDied","Data":"6f3c01688bba973a49cd38ead6f32071203a13f1c437e87590a2062f43c42d3e"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.552406 4907 scope.go:117] "RemoveContainer" containerID="3fd8ece48af6f3186fe1a67ce35dde9ab752271f826eb2d327ed3e6092299012" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.552506 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f88567fd9-f24rf" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.559129 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4tskm" podStartSLOduration=2.559112485 podStartE2EDuration="2.559112485s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:58.552384451 +0000 UTC m=+1317.452172140" watchObservedRunningTime="2026-03-13 14:26:58.559112485 +0000 UTC m=+1317.458900174" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.560334 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.572869 4907 generic.go:334] "Generic (PLEG): container finished" podID="c495e208-654b-477a-869e-87a5c2566519" containerID="12156f70b991f6a7b1a9f418d96256568ed61931e620f9a12a50e33bb305271b" exitCode=0 Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.572996 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" event={"ID":"c495e208-654b-477a-869e-87a5c2566519","Type":"ContainerDied","Data":"12156f70b991f6a7b1a9f418d96256568ed61931e620f9a12a50e33bb305271b"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.573126 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" event={"ID":"c495e208-654b-477a-869e-87a5c2566519","Type":"ContainerStarted","Data":"6116fd540adb884533031b5d7e277d5df41a49b6c0e586ce5363f34faed7d7b7"} Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.591435 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.624714 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.624993 4907 scope.go:117] "RemoveContainer" containerID="b20445b78457705b4b60b2201ff58c281da6086f771d56be4ceb3ab98dc8b1f5" Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.633450 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f88567fd9-f24rf"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.717171 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.728309 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:26:58 crc kubenswrapper[4907]: W0313 14:26:58.729410 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32b531c6_64c1_4137_b82e_3c10789325e6.slice/crio-4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929 WatchSource:0}: Error finding container 4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929: Status 404 returned error can't find the container with id 4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929 Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.735643 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.763229 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:26:58 crc kubenswrapper[4907]: I0313 14:26:58.789651 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:26:58 crc kubenswrapper[4907]: W0313 14:26:58.821513 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda661b9e1_c551_4b36_87a9_833df73bd6b2.slice/crio-eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0 WatchSource:0}: Error finding container eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0: Status 404 returned error can't find the container with id eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0 Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.054079 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.182589 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227346 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227478 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227783 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227829 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.227951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") pod \"c495e208-654b-477a-869e-87a5c2566519\" (UID: \"c495e208-654b-477a-869e-87a5c2566519\") " Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.235472 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278" (OuterVolumeSpecName: "kube-api-access-db278") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "kube-api-access-db278". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.247798 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.278531 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.281279 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.308535 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config" (OuterVolumeSpecName: "config") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.314648 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.351661 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357437 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357478 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357491 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357503 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357514 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-db278\" (UniqueName: \"kubernetes.io/projected/c495e208-654b-477a-869e-87a5c2566519-kube-api-access-db278\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.357581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c495e208-654b-477a-869e-87a5c2566519" (UID: "c495e208-654b-477a-869e-87a5c2566519"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.436692 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.460395 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c495e208-654b-477a-869e-87a5c2566519-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.586341 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerStarted","Data":"e365b5ca2986747d106a442f14cd719005e64c8ba4070ee9236f7119ccd3db0a"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.589720 4907 generic.go:334] "Generic (PLEG): container finished" podID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerID="4e3ebec78f62227a1eec084cd7d40d9bb86b61261221bf385169185e602cfad5" exitCode=0 Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.589844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerDied","Data":"4e3ebec78f62227a1eec084cd7d40d9bb86b61261221bf385169185e602cfad5"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.589910 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerStarted","Data":"eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.596380 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerStarted","Data":"c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.596426 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerStarted","Data":"61484e5c865e06dbc491f0726df5e379526c160f176a2435939232b317293ca1"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.609637 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" event={"ID":"c495e208-654b-477a-869e-87a5c2566519","Type":"ContainerDied","Data":"6116fd540adb884533031b5d7e277d5df41a49b6c0e586ce5363f34faed7d7b7"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.609690 4907 scope.go:117] "RemoveContainer" containerID="12156f70b991f6a7b1a9f418d96256568ed61931e620f9a12a50e33bb305271b" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.609796 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.616918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerStarted","Data":"4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.637543 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerStarted","Data":"247e75fb8f70b076f7d10c60ec2902246989846bb77cff82137c6d0beb413598"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.640578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerStarted","Data":"d144653276b0a0015c06098911900d038b6dd5b20a40a40339bc3b434d2953a5"} Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.644770 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jt88g" podStartSLOduration=2.644756852 podStartE2EDuration="2.644756852s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:26:59.631989244 +0000 UTC m=+1318.531776933" watchObservedRunningTime="2026-03-13 14:26:59.644756852 +0000 UTC m=+1318.544544541" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.692624 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.705187 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cb4dcfdd7-k6pkx"] Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.939224 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aba6c70-a80a-4bce-9216-887cb7a2061c" path="/var/lib/kubelet/pods/7aba6c70-a80a-4bce-9216-887cb7a2061c/volumes" Mar 13 14:26:59 crc kubenswrapper[4907]: I0313 14:26:59.940058 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c495e208-654b-477a-869e-87a5c2566519" path="/var/lib/kubelet/pods/c495e208-654b-477a-869e-87a5c2566519/volumes" Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.319757 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.683432 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerStarted","Data":"70f4055057ae74811d3e519a30dbdf9559ebe320aa1f7b52ebbfd8cfd47c2673"} Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.686960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerStarted","Data":"d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f"} Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.699661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerStarted","Data":"46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4"} Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.700505 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:00 crc kubenswrapper[4907]: I0313 14:27:00.730253 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" podStartSLOduration=3.730231944 podStartE2EDuration="3.730231944s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:00.724065206 +0000 UTC m=+1319.623852895" watchObservedRunningTime="2026-03-13 14:27:00.730231944 +0000 UTC m=+1319.630019633" Mar 13 14:27:01 crc kubenswrapper[4907]: I0313 14:27:01.739419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerStarted","Data":"9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1"} Mar 13 14:27:01 crc kubenswrapper[4907]: I0313 14:27:01.743302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerStarted","Data":"512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778"} Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.754227 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerStarted","Data":"c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149"} Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.754307 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" containerID="cri-o://d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.755128 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" containerID="cri-o://512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.755240 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" containerID="cri-o://c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.755331 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" containerID="cri-o://9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1" gracePeriod=30 Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.784073 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.784019912 podStartE2EDuration="6.784019912s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:02.778303467 +0000 UTC m=+1321.678091156" watchObservedRunningTime="2026-03-13 14:27:02.784019912 +0000 UTC m=+1321.683807601" Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.804058 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.804042417 podStartE2EDuration="6.804042417s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:02.800962223 +0000 UTC m=+1321.700749912" watchObservedRunningTime="2026-03-13 14:27:02.804042417 +0000 UTC m=+1321.703830106" Mar 13 14:27:02 crc kubenswrapper[4907]: I0313 14:27:02.897865 4907 scope.go:117] "RemoveContainer" containerID="dd6efe3aa3a79d161b96b8d84e7fa19f50f23cdeadb3268906a3667faac57368" Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.764416 4907 generic.go:334] "Generic (PLEG): container finished" podID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerID="8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f" exitCode=0 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.764524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerDied","Data":"8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766897 4907 generic.go:334] "Generic (PLEG): container finished" podID="ce2a993a-c706-400f-967d-35a9c38d5937" containerID="c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149" exitCode=0 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766915 4907 generic.go:334] "Generic (PLEG): container finished" podID="ce2a993a-c706-400f-967d-35a9c38d5937" containerID="512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778" exitCode=143 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766922 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerDied","Data":"c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.766957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerDied","Data":"512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769807 4907 generic.go:334] "Generic (PLEG): container finished" podID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerID="9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1" exitCode=0 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769827 4907 generic.go:334] "Generic (PLEG): container finished" podID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerID="d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f" exitCode=143 Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerDied","Data":"9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1"} Mar 13 14:27:03 crc kubenswrapper[4907]: I0313 14:27:03.769862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerDied","Data":"d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f"} Mar 13 14:27:07 crc kubenswrapper[4907]: I0313 14:27:07.756146 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:07 crc kubenswrapper[4907]: I0313 14:27:07.848226 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:27:07 crc kubenswrapper[4907]: I0313 14:27:07.848506 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67754df655-k2v49" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" containerID="cri-o://fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213" gracePeriod=10 Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.847669 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4tskm" event={"ID":"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4","Type":"ContainerDied","Data":"a2858c30fa2ae4911b2a7afe4d6fb52407d5d3328c466096a4d4cc1761524119"} Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.848096 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2858c30fa2ae4911b2a7afe4d6fb52407d5d3328c466096a4d4cc1761524119" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.850616 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.851008 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5","Type":"ContainerDied","Data":"e365b5ca2986747d106a442f14cd719005e64c8ba4070ee9236f7119ccd3db0a"} Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.851081 4907 scope.go:117] "RemoveContainer" containerID="9d6a810392296b49581d1d82cea96d2320cf1660821d93127c921446e01396f1" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.854642 4907 generic.go:334] "Generic (PLEG): container finished" podID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerID="fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213" exitCode=0 Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.854715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerDied","Data":"fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213"} Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.859456 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963012 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963086 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963109 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963125 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963143 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963163 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963183 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963198 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963218 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963236 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963265 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") pod \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\" (UID: \"4a97d1cc-3762-47fe-a8fe-ed9e95b466a4\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963380 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.963414 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") pod \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\" (UID: \"8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5\") " Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.966285 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs" (OuterVolumeSpecName: "logs") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.966516 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.971692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts" (OuterVolumeSpecName: "scripts") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.971968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp" (OuterVolumeSpecName: "kube-api-access-k48mp") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "kube-api-access-k48mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.972512 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.973017 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.973152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq" (OuterVolumeSpecName: "kube-api-access-pvhgq") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "kube-api-access-pvhgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.973513 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts" (OuterVolumeSpecName: "scripts") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.977574 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.997562 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:08 crc kubenswrapper[4907]: I0313 14:27:08.998132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data" (OuterVolumeSpecName: "config-data") pod "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" (UID: "4a97d1cc-3762-47fe-a8fe-ed9e95b466a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.011022 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.036663 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.036997 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data" (OuterVolumeSpecName: "config-data") pod "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" (UID: "8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.065979 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066018 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066031 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066145 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvhgq\" (UniqueName: \"kubernetes.io/projected/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-kube-api-access-pvhgq\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066157 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066167 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066178 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066189 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066227 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066239 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066305 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066320 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k48mp\" (UniqueName: \"kubernetes.io/projected/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-kube-api-access-k48mp\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066329 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.066338 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.095700 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.168174 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.719648 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67754df655-k2v49" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.869399 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4tskm" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.869454 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.894396 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.903872 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951083 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951503 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c495e208-654b-477a-869e-87a5c2566519" containerName="init" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951524 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c495e208-654b-477a-869e-87a5c2566519" containerName="init" Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951542 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951550 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951565 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerName="keystone-bootstrap" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerName="keystone-bootstrap" Mar 13 14:27:09 crc kubenswrapper[4907]: E0313 14:27:09.951600 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951608 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951788 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" containerName="keystone-bootstrap" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951798 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c495e208-654b-477a-869e-87a5c2566519" containerName="init" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951807 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-log" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.951828 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" containerName="glance-httpd" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.952954 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.955450 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.955829 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.975445 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.986298 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:27:09 crc kubenswrapper[4907]: I0313 14:27:09.994816 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4tskm"] Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.056058 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.057127 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.059019 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.059124 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.059304 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.060841 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.062961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.074569 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092842 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092913 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092936 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.092981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.093015 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.093038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.093061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195001 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195131 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195157 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195187 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195239 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195292 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195372 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195494 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.195539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.296975 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297168 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.297254 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.304433 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.304539 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.304546 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.306014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.307173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.317388 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"keystone-bootstrap-crrww\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.323409 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.323450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.323627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.325190 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.327142 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.329361 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.330520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.354987 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.413990 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:10 crc kubenswrapper[4907]: I0313 14:27:10.572594 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.638701 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719787 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719839 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719908 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719926 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.719962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.720043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.720062 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ce2a993a-c706-400f-967d-35a9c38d5937\" (UID: \"ce2a993a-c706-400f-967d-35a9c38d5937\") " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.721382 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs" (OuterVolumeSpecName: "logs") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.721816 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.725937 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26" (OuterVolumeSpecName: "kube-api-access-jxx26") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "kube-api-access-jxx26". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.726028 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts" (OuterVolumeSpecName: "scripts") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.754901 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.772190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data" (OuterVolumeSpecName: "config-data") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.773081 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.787112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ce2a993a-c706-400f-967d-35a9c38d5937" (UID: "ce2a993a-c706-400f-967d-35a9c38d5937"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.794226 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a97d1cc-3762-47fe-a8fe-ed9e95b466a4" path="/var/lib/kubelet/pods/4a97d1cc-3762-47fe-a8fe-ed9e95b466a4/volumes" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.795213 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5" path="/var/lib/kubelet/pods/8d198bcd-d3e5-43e3-bd97-8d6e7f11b4c5/volumes" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822736 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822798 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822811 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822825 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822836 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822846 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce2a993a-c706-400f-967d-35a9c38d5937-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822856 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxx26\" (UniqueName: \"kubernetes.io/projected/ce2a993a-c706-400f-967d-35a9c38d5937-kube-api-access-jxx26\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.822867 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce2a993a-c706-400f-967d-35a9c38d5937-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.843522 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.899519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce2a993a-c706-400f-967d-35a9c38d5937","Type":"ContainerDied","Data":"70f4055057ae74811d3e519a30dbdf9559ebe320aa1f7b52ebbfd8cfd47c2673"} Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.899620 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.924129 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.924530 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.936983 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983134 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:11 crc kubenswrapper[4907]: E0313 14:27:11.983575 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983597 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" Mar 13 14:27:11 crc kubenswrapper[4907]: E0313 14:27:11.983639 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983648 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983813 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-httpd" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.983843 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" containerName="glance-log" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.984743 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.987456 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.987604 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 14:27:11 crc kubenswrapper[4907]: I0313 14:27:11.992105 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.129853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.129927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.129984 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.130431 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232593 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232717 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232740 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232777 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232820 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232850 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.232939 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.235864 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.236022 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.236076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.236587 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.238453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.240179 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.243456 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.247040 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.249267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.262295 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " pod="openstack/glance-default-external-api-0" Mar 13 14:27:12 crc kubenswrapper[4907]: I0313 14:27:12.305651 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:27:13 crc kubenswrapper[4907]: I0313 14:27:13.791521 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce2a993a-c706-400f-967d-35a9c38d5937" path="/var/lib/kubelet/pods/ce2a993a-c706-400f-967d-35a9c38d5937/volumes" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.662997 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750170 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750241 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750315 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750400 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750461 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.750497 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") pod \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\" (UID: \"c065ce46-d7f3-4e50-b30c-83a1592b6ead\") " Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.756040 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g" (OuterVolumeSpecName: "kube-api-access-vpx9g") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "kube-api-access-vpx9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.800904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.809026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config" (OuterVolumeSpecName: "config") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.809042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.813361 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.815247 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c065ce46-d7f3-4e50-b30c-83a1592b6ead" (UID: "c065ce46-d7f3-4e50-b30c-83a1592b6ead"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853168 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853223 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853234 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853289 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853321 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c065ce46-d7f3-4e50-b30c-83a1592b6ead-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.853335 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpx9g\" (UniqueName: \"kubernetes.io/projected/c065ce46-d7f3-4e50-b30c-83a1592b6ead-kube-api-access-vpx9g\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.969816 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67754df655-k2v49" event={"ID":"c065ce46-d7f3-4e50-b30c-83a1592b6ead","Type":"ContainerDied","Data":"1dc89b2197970bff5fe57426501341e047b5f1d0b4ba93884702cc3864ea9041"} Mar 13 14:27:18 crc kubenswrapper[4907]: I0313 14:27:18.969953 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67754df655-k2v49" Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.003112 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.011413 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67754df655-k2v49"] Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.719549 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-67754df655-k2v49" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Mar 13 14:27:19 crc kubenswrapper[4907]: I0313 14:27:19.792292 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" path="/var/lib/kubelet/pods/c065ce46-d7f3-4e50-b30c-83a1592b6ead/volumes" Mar 13 14:27:20 crc kubenswrapper[4907]: I0313 14:27:20.931463 4907 scope.go:117] "RemoveContainer" containerID="d56b748bd6482c49c5b182589b671f235af50f54159cfd34d117c3100423cb0f" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.964361 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.964497 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wpft2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7q999_openstack(594b61fa-12b7-47b6-8af3-17f024e0d54d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.965644 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7q999" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" Mar 13 14:27:20 crc kubenswrapper[4907]: E0313 14:27:20.994704 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:574a17f0877c175128a764f2b37fc02456649c8514689125718ce6ca974bfb6b\\\"\"" pod="openstack/cinder-db-sync-7q999" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.146806 4907 scope.go:117] "RemoveContainer" containerID="c604cb7d84d59e89fd67f7b2707cc5e5fce549a56d035662c24b9cb3f930d149" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.201298 4907 scope.go:117] "RemoveContainer" containerID="512291676abb06ed48b9f163f1a7202f5fd9b599abc8f6b456c67f77c909e778" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.250705 4907 scope.go:117] "RemoveContainer" containerID="fd0a2c793e62b5b128e16861d049a22d36d9baa0de3410f3961ca59044ad1213" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.286679 4907 scope.go:117] "RemoveContainer" containerID="fc5cca7a53563949cc0a2a52bb6bc5ab553013ec2df261991d02d7a2bd1a082d" Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.412607 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:27:21 crc kubenswrapper[4907]: W0313 14:27:21.415533 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34a0ea40_bdaa_4eef_8a28_263d8db74f10.slice/crio-35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb WatchSource:0}: Error finding container 35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb: Status 404 returned error can't find the container with id 35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.586947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:27:21 crc kubenswrapper[4907]: W0313 14:27:21.589063 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c36a352_151f_4e93_8094_3855bfed532e.slice/crio-4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23 WatchSource:0}: Error finding container 4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23: Status 404 returned error can't find the container with id 4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23 Mar 13 14:27:21 crc kubenswrapper[4907]: I0313 14:27:21.692018 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.008646 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerStarted","Data":"5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.015436 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerStarted","Data":"70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.027532 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerStarted","Data":"f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.027573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerStarted","Data":"35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.047367 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.050407 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-rz6qh" podStartSLOduration=2.827143362 podStartE2EDuration="25.050385457s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.736334635 +0000 UTC m=+1317.636122324" lastFinishedPulling="2026-03-13 14:27:20.95957673 +0000 UTC m=+1339.859364419" observedRunningTime="2026-03-13 14:27:22.040599181 +0000 UTC m=+1340.940386870" watchObservedRunningTime="2026-03-13 14:27:22.050385457 +0000 UTC m=+1340.950173156" Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.050602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerStarted","Data":"2cde1712ac814253467a7ebeac79ad28f1b6bc8f06f7979a4876a959f01886bd"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.051945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerStarted","Data":"4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23"} Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.065713 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-crrww" podStartSLOduration=12.065692364 podStartE2EDuration="12.065692364s" podCreationTimestamp="2026-03-13 14:27:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:22.06115099 +0000 UTC m=+1340.960938679" watchObservedRunningTime="2026-03-13 14:27:22.065692364 +0000 UTC m=+1340.965480053" Mar 13 14:27:22 crc kubenswrapper[4907]: I0313 14:27:22.080796 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-jjqr9" podStartSLOduration=2.953097157 podStartE2EDuration="25.080778144s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.808536038 +0000 UTC m=+1317.708323727" lastFinishedPulling="2026-03-13 14:27:20.936217025 +0000 UTC m=+1339.836004714" observedRunningTime="2026-03-13 14:27:22.078459401 +0000 UTC m=+1340.978247090" watchObservedRunningTime="2026-03-13 14:27:22.080778144 +0000 UTC m=+1340.980565833" Mar 13 14:27:23 crc kubenswrapper[4907]: I0313 14:27:23.067209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512"} Mar 13 14:27:23 crc kubenswrapper[4907]: I0313 14:27:23.070969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerStarted","Data":"6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5"} Mar 13 14:27:23 crc kubenswrapper[4907]: I0313 14:27:23.074454 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerStarted","Data":"4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450"} Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.084566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerStarted","Data":"22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f"} Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.087193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerStarted","Data":"456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f"} Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.131336 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=15.131318393 podStartE2EDuration="15.131318393s" podCreationTimestamp="2026-03-13 14:27:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:24.13011225 +0000 UTC m=+1343.029899939" watchObservedRunningTime="2026-03-13 14:27:24.131318393 +0000 UTC m=+1343.031106082" Mar 13 14:27:24 crc kubenswrapper[4907]: I0313 14:27:24.131700 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=13.131695053 podStartE2EDuration="13.131695053s" podCreationTimestamp="2026-03-13 14:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:24.106489078 +0000 UTC m=+1343.006276767" watchObservedRunningTime="2026-03-13 14:27:24.131695053 +0000 UTC m=+1343.031482742" Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.112003 4907 generic.go:334] "Generic (PLEG): container finished" podID="32b531c6-64c1-4137-b82e-3c10789325e6" containerID="5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034" exitCode=0 Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.112183 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerDied","Data":"5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034"} Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.117089 4907 generic.go:334] "Generic (PLEG): container finished" podID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerID="70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa" exitCode=0 Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.117169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerDied","Data":"70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa"} Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.119875 4907 generic.go:334] "Generic (PLEG): container finished" podID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerID="f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5" exitCode=0 Mar 13 14:27:25 crc kubenswrapper[4907]: I0313 14:27:25.121255 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerDied","Data":"f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5"} Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.130716 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b"} Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.577135 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.582952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.588626 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688383 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688522 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688572 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688606 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688643 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688694 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688723 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") pod \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\" (UID: \"34a0ea40-bdaa-4eef-8a28-263d8db74f10\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688917 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") pod \"b144897c-8640-4c5c-afa1-68b44dfd4f98\" (UID: \"b144897c-8640-4c5c-afa1-68b44dfd4f98\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.688943 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") pod \"32b531c6-64c1-4137-b82e-3c10789325e6\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.690298 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs" (OuterVolumeSpecName: "logs") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.695685 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.695733 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "32b531c6-64c1-4137-b82e-3c10789325e6" (UID: "32b531c6-64c1-4137-b82e-3c10789325e6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.695802 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67" (OuterVolumeSpecName: "kube-api-access-wcv67") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "kube-api-access-wcv67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.696117 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts" (OuterVolumeSpecName: "scripts") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.696470 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt" (OuterVolumeSpecName: "kube-api-access-8grrt") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "kube-api-access-8grrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.696556 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts" (OuterVolumeSpecName: "scripts") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.697384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.717042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.720092 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.722012 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data" (OuterVolumeSpecName: "config-data") pod "b144897c-8640-4c5c-afa1-68b44dfd4f98" (UID: "b144897c-8640-4c5c-afa1-68b44dfd4f98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.722171 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data" (OuterVolumeSpecName: "config-data") pod "34a0ea40-bdaa-4eef-8a28-263d8db74f10" (UID: "34a0ea40-bdaa-4eef-8a28-263d8db74f10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.791432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") pod \"32b531c6-64c1-4137-b82e-3c10789325e6\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.791596 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") pod \"32b531c6-64c1-4137-b82e-3c10789325e6\" (UID: \"32b531c6-64c1-4137-b82e-3c10789325e6\") " Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792104 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792123 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792137 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792149 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792161 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.792173 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b144897c-8640-4c5c-afa1-68b44dfd4f98-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793241 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793258 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8grrt\" (UniqueName: \"kubernetes.io/projected/34a0ea40-bdaa-4eef-8a28-263d8db74f10-kube-api-access-8grrt\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793270 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b144897c-8640-4c5c-afa1-68b44dfd4f98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793281 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34a0ea40-bdaa-4eef-8a28-263d8db74f10-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcv67\" (UniqueName: \"kubernetes.io/projected/b144897c-8640-4c5c-afa1-68b44dfd4f98-kube-api-access-wcv67\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.793307 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.802411 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc" (OuterVolumeSpecName: "kube-api-access-578zc") pod "32b531c6-64c1-4137-b82e-3c10789325e6" (UID: "32b531c6-64c1-4137-b82e-3c10789325e6"). InnerVolumeSpecName "kube-api-access-578zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.823647 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32b531c6-64c1-4137-b82e-3c10789325e6" (UID: "32b531c6-64c1-4137-b82e-3c10789325e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.895101 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-578zc\" (UniqueName: \"kubernetes.io/projected/32b531c6-64c1-4137-b82e-3c10789325e6-kube-api-access-578zc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:26 crc kubenswrapper[4907]: I0313 14:27:26.895140 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32b531c6-64c1-4137-b82e-3c10789325e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.143748 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-rz6qh" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.143747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-rz6qh" event={"ID":"32b531c6-64c1-4137-b82e-3c10789325e6","Type":"ContainerDied","Data":"4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.144222 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b7673a5ff0b4e402bcd5b85ece6329b6bba3e9faaac7ef9c519ffe96fec5929" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.145565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jjqr9" event={"ID":"b144897c-8640-4c5c-afa1-68b44dfd4f98","Type":"ContainerDied","Data":"d144653276b0a0015c06098911900d038b6dd5b20a40a40339bc3b434d2953a5"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.145597 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d144653276b0a0015c06098911900d038b6dd5b20a40a40339bc3b434d2953a5" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.145612 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jjqr9" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.151161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-crrww" event={"ID":"34a0ea40-bdaa-4eef-8a28-263d8db74f10","Type":"ContainerDied","Data":"35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.151198 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35ce55e83a38917576782260db545e824c3f1f351d75a47429e1e2d417a299bb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.151318 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-crrww" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.153425 4907 generic.go:334] "Generic (PLEG): container finished" podID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerID="c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb" exitCode=0 Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.153455 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerDied","Data":"c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb"} Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318468 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.318896 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerName="placement-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318919 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerName="placement-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.318945 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerName="keystone-bootstrap" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318953 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerName="keystone-bootstrap" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.318975 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" containerName="barbican-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.318985 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" containerName="barbican-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.319005 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319014 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" Mar 13 14:27:27 crc kubenswrapper[4907]: E0313 14:27:27.319033 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="init" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319041 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="init" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319250 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" containerName="keystone-bootstrap" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319276 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" containerName="barbican-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319293 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" containerName="placement-db-sync" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.319308 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c065ce46-d7f3-4e50-b30c-83a1592b6ead" containerName="dnsmasq-dns" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.320041 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325648 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325917 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325937 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.325989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.326144 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.326323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-p9m6m" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.342960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.356127 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.364907 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.374390 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pg28b" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377331 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377553 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377696 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.377943 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.409382 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.419830 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.430562 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.458648 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-skqjb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.458866 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.520964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521039 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521076 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521125 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521225 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521371 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521470 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.521529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.531822 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.541600 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.549014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.551925 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.576394 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.596522 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626771 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626831 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626850 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626876 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626915 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.626999 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627018 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627040 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627060 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627082 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627102 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627123 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627147 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627163 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.630069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.627856 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.635643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.636440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.638674 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.639274 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.640633 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.640679 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.644610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.645015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.645523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.648387 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.648826 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.651449 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.657398 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.663149 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"placement-67d6ccd9c4-ms98l\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.664673 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"keystone-85b9d64669-dxnf6\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.670094 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.710048 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.713473 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.725868 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.726045 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.728943 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.728987 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729041 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729069 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729135 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729156 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729175 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729198 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729266 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729289 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.729348 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.735250 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.740856 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.741633 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.742227 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.743590 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.793806 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"barbican-worker-c685c5475-rvkf2\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832785 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832812 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832833 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832925 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832950 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.832981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833004 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833071 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833125 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.833482 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.837718 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.838069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.842040 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.842548 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.843292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.844144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.844943 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.871394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.874740 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.880333 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"barbican-keystone-listener-54c69f4bdb-lt67r\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.881167 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.881249 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.905538 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.907038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.917141 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.945142 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.954824 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"dnsmasq-dns-79fd7f986f-t9wlb\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961515 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961691 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961784 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.961846 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.965310 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.981466 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.982194 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:27 crc kubenswrapper[4907]: I0313 14:27:27.982702 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.005825 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.025748 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"barbican-api-5d49fd688d-njvfm\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.043586 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.053280 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.054613 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063712 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063740 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063774 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063848 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063960 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.063986 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.064009 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.064038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.064077 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.083516 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.094551 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165417 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165842 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165864 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165946 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165966 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.165993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166029 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166087 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166111 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166159 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166182 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166208 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166224 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.166252 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.170916 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.171759 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.172010 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.176611 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.177039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.185557 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.186894 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.215475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.218043 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"barbican-keystone-listener-7d7bc9f7d-fks76\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.218625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.222542 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"barbican-worker-76bbcdf7d9-j6hw5\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267730 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267860 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267955 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.267983 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.269382 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.274538 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.281388 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.289167 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.290015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"barbican-api-74984bd8b4-x42pq\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.393323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.425418 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.438096 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.897100 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.924251 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:27:28 crc kubenswrapper[4907]: W0313 14:27:28.932960 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a738a66_54bc_4f5e_86d5_6e3004d8f265.slice/crio-50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4 WatchSource:0}: Error finding container 50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4: Status 404 returned error can't find the container with id 50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4 Mar 13 14:27:28 crc kubenswrapper[4907]: I0313 14:27:28.963517 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:29 crc kubenswrapper[4907]: W0313 14:27:28.995522 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca4928ba_1277_4918_8b92_3eda2b276577.slice/crio-6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2 WatchSource:0}: Error finding container 6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2: Status 404 returned error can't find the container with id 6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2 Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.095251 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.105843 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:29 crc kubenswrapper[4907]: W0313 14:27:29.115042 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d754127_d8df_48de_b77a_46aa7898c45f.slice/crio-90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50 WatchSource:0}: Error finding container 90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50: Status 404 returned error can't find the container with id 90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50 Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.117315 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.123226 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:29 crc kubenswrapper[4907]: W0313 14:27:29.130112 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod070e2b75_05b1_42a4_a34e_950d45d65920.slice/crio-eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7 WatchSource:0}: Error finding container eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7: Status 404 returned error can't find the container with id eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7 Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.203172 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") pod \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.203277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") pod \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.210550 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t" (OuterVolumeSpecName: "kube-api-access-wq68t") pod "870ec55c-fc6f-4b16-8c49-a3cda0d0d010" (UID: "870ec55c-fc6f-4b16-8c49-a3cda0d0d010"). InnerVolumeSpecName "kube-api-access-wq68t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.284302 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config" (OuterVolumeSpecName: "config") pod "870ec55c-fc6f-4b16-8c49-a3cda0d0d010" (UID: "870ec55c-fc6f-4b16-8c49-a3cda0d0d010"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.289918 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.310415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.311976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") pod \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\" (UID: \"870ec55c-fc6f-4b16-8c49-a3cda0d0d010\") " Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.312650 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wq68t\" (UniqueName: \"kubernetes.io/projected/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-kube-api-access-wq68t\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.312663 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.321169 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.333390 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jt88g" event={"ID":"870ec55c-fc6f-4b16-8c49-a3cda0d0d010","Type":"ContainerDied","Data":"61484e5c865e06dbc491f0726df5e379526c160f176a2435939232b317293ca1"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.333443 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61484e5c865e06dbc491f0726df5e379526c160f176a2435939232b317293ca1" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.333510 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jt88g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.356586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerStarted","Data":"50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.381361 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerStarted","Data":"7a66d01a94d06b4565675b1fcb3708a34133ebf672d54a557137e862ae4fd1c3"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.399193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerStarted","Data":"6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.413042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "870ec55c-fc6f-4b16-8c49-a3cda0d0d010" (UID: "870ec55c-fc6f-4b16-8c49-a3cda0d0d010"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.415317 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/870ec55c-fc6f-4b16-8c49-a3cda0d0d010-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.434294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerStarted","Data":"eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.438011 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.445159 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerStarted","Data":"e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.445206 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerStarted","Data":"0497ac4d8503d6e45cbe04af5eff7d939b60b6a6232917e037fb11fdbb11b8dd"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.476629 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" event={"ID":"2d754127-d8df-48de-b77a-46aa7898c45f","Type":"ContainerStarted","Data":"90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50"} Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.480854 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:29 crc kubenswrapper[4907]: E0313 14:27:29.481297 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerName="neutron-db-sync" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.481310 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerName="neutron-db-sync" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.481497 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" containerName="neutron-db-sync" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.482472 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519008 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519051 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519127 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519158 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519280 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.519322 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.545127 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621011 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621056 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621081 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621137 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621236 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.621972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.624976 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.625604 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.626467 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.629516 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.663267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"dnsmasq-dns-7fc46d7df7-77dng\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.677638 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.679331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.687757 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-tvcdf" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.688072 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.688226 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.688843 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.731740 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.833453 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.834423 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.834565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.834775 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.835778 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.871221 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939580 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939681 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939701 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939799 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.939925 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.946316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.948589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.951991 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.953615 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:29 crc kubenswrapper[4907]: I0313 14:27:29.974439 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"neutron-944cbc5bb-fj29g\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.023923 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.330441 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.494720 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerStarted","Data":"38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.495388 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.495418 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.501366 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerStarted","Data":"57bab78a65adf0bc133213358e1f66cc1c97a4b0aefeb088fc087562e2ca2999"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.504379 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerStarted","Data":"74e19d820a0370c065d553cb0dbb43ed632fe5214db6fe0a8a5f25c3d63f84ae"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.507085 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d754127-d8df-48de-b77a-46aa7898c45f" containerID="9ea73af353594fa5d86ca1191ad584d04081c81707270b01c9c5415f0d4cc909" exitCode=0 Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.507154 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" event={"ID":"2d754127-d8df-48de-b77a-46aa7898c45f","Type":"ContainerDied","Data":"9ea73af353594fa5d86ca1191ad584d04081c81707270b01c9c5415f0d4cc909"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.508974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerStarted","Data":"be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.509009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerStarted","Data":"3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.509023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerStarted","Data":"3fa9f3f934d41bda9d2c0367bd05516b9fcfddd3de52bfe4494f0529db4eafd7"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.510392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerStarted","Data":"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.510538 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.511623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerStarted","Data":"d83bd2c5278d976f4863dc597e16a40e27298e4624048379804b2a946c5145cc"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.513153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerStarted","Data":"8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.513199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerStarted","Data":"57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624"} Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.513311 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.539862 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-67d6ccd9c4-ms98l" podStartSLOduration=3.539841278 podStartE2EDuration="3.539841278s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:30.525780136 +0000 UTC m=+1349.425567825" watchObservedRunningTime="2026-03-13 14:27:30.539841278 +0000 UTC m=+1349.439628967" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.572461 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-85b9d64669-dxnf6" podStartSLOduration=3.572435484 podStartE2EDuration="3.572435484s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:30.562618588 +0000 UTC m=+1349.462406277" watchObservedRunningTime="2026-03-13 14:27:30.572435484 +0000 UTC m=+1349.472223173" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.574415 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.574469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.608818 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d49fd688d-njvfm" podStartSLOduration=3.608799904 podStartE2EDuration="3.608799904s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:30.597497176 +0000 UTC m=+1349.497284875" watchObservedRunningTime="2026-03-13 14:27:30.608799904 +0000 UTC m=+1349.508587593" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.660298 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.660407 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.804396 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:27:30 crc kubenswrapper[4907]: I0313 14:27:30.972858 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.068660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.068819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.068870 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.071085 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.071205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.071319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") pod \"2d754127-d8df-48de-b77a-46aa7898c45f\" (UID: \"2d754127-d8df-48de-b77a-46aa7898c45f\") " Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.079035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg" (OuterVolumeSpecName: "kube-api-access-dhzwg") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "kube-api-access-dhzwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.119374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config" (OuterVolumeSpecName: "config") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.133979 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.134291 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.132808 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.159772 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2d754127-d8df-48de-b77a-46aa7898c45f" (UID: "2d754127-d8df-48de-b77a-46aa7898c45f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177356 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177398 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177413 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177424 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177437 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhzwg\" (UniqueName: \"kubernetes.io/projected/2d754127-d8df-48de-b77a-46aa7898c45f-kube-api-access-dhzwg\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.177449 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d754127-d8df-48de-b77a-46aa7898c45f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.522477 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerStarted","Data":"b6a3399278c9114cdb41d239ecf9266ca04769841425ab359d3ec15a0ea2369a"} Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.527637 4907 generic.go:334] "Generic (PLEG): container finished" podID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerID="577048fe225fd9933f67497a9e39e7a61a8a9a2ce395fd1bd05afbdcf17490ca" exitCode=0 Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.527684 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerDied","Data":"577048fe225fd9933f67497a9e39e7a61a8a9a2ce395fd1bd05afbdcf17490ca"} Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.530376 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.530419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79fd7f986f-t9wlb" event={"ID":"2d754127-d8df-48de-b77a-46aa7898c45f","Type":"ContainerDied","Data":"90e9289f409305f490b74370d98aff58f19c0c80ed8fdee54b693082f3b3ca50"} Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.530484 4907 scope.go:117] "RemoveContainer" containerID="9ea73af353594fa5d86ca1191ad584d04081c81707270b01c9c5415f0d4cc909" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.531432 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.531469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.531481 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.580238 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-74984bd8b4-x42pq" podStartSLOduration=4.580218064 podStartE2EDuration="4.580218064s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:31.566056919 +0000 UTC m=+1350.465844628" watchObservedRunningTime="2026-03-13 14:27:31.580218064 +0000 UTC m=+1350.480005753" Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.629926 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.640608 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79fd7f986f-t9wlb"] Mar 13 14:27:31 crc kubenswrapper[4907]: I0313 14:27:31.800589 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" path="/var/lib/kubelet/pods/2d754127-d8df-48de-b77a-46aa7898c45f/volumes" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.308257 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.308650 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.350037 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.365583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.552716 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:27:32 crc kubenswrapper[4907]: I0313 14:27:32.552746 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.438430 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.438979 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.571470 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerStarted","Data":"10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.571545 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.575450 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerStarted","Data":"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.577610 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerStarted","Data":"6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.579294 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerStarted","Data":"bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.580930 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerStarted","Data":"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.582981 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerStarted","Data":"b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.583022 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerStarted","Data":"a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75"} Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.583307 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.601425 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" podStartSLOduration=4.601405385 podStartE2EDuration="4.601405385s" podCreationTimestamp="2026-03-13 14:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:33.595768522 +0000 UTC m=+1352.495556211" watchObservedRunningTime="2026-03-13 14:27:33.601405385 +0000 UTC m=+1352.501193074" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.653440 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-944cbc5bb-fj29g" podStartSLOduration=4.65341697 podStartE2EDuration="4.65341697s" podCreationTimestamp="2026-03-13 14:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:33.617595455 +0000 UTC m=+1352.517383144" watchObservedRunningTime="2026-03-13 14:27:33.65341697 +0000 UTC m=+1352.553204659" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.841893 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:27:33 crc kubenswrapper[4907]: E0313 14:27:33.842267 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" containerName="init" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.842284 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" containerName="init" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.842472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d754127-d8df-48de-b77a-46aa7898c45f" containerName="init" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.843406 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.846593 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.846842 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.860979 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941733 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941826 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941942 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.941987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.942162 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:33 crc kubenswrapper[4907]: I0313 14:27:33.942270 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.044911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045341 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045385 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.045420 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.047968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.048070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.053678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.054235 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.054643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.055760 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.059716 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.077725 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.091723 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.092107 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.093473 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"neutron-7bf587b8d9-pvvbz\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.189133 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.604971 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerStarted","Data":"9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.629218 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerStarted","Data":"2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.637367 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerStarted","Data":"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.675422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerStarted","Data":"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4"} Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.703502 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" podStartSLOduration=4.525446015 podStartE2EDuration="7.703478959s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:29.451100348 +0000 UTC m=+1348.350888037" lastFinishedPulling="2026-03-13 14:27:32.629133292 +0000 UTC m=+1351.528920981" observedRunningTime="2026-03-13 14:27:34.699736307 +0000 UTC m=+1353.599524016" watchObservedRunningTime="2026-03-13 14:27:34.703478959 +0000 UTC m=+1353.603266648" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.718233 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" podStartSLOduration=4.201840112 podStartE2EDuration="7.718209029s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:29.100366668 +0000 UTC m=+1348.000154357" lastFinishedPulling="2026-03-13 14:27:32.616735585 +0000 UTC m=+1351.516523274" observedRunningTime="2026-03-13 14:27:34.652292496 +0000 UTC m=+1353.552080185" watchObservedRunningTime="2026-03-13 14:27:34.718209029 +0000 UTC m=+1353.617996718" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.751843 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-c685c5475-rvkf2" podStartSLOduration=4.11569887 podStartE2EDuration="7.751821763s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:28.999928067 +0000 UTC m=+1347.899715756" lastFinishedPulling="2026-03-13 14:27:32.63605096 +0000 UTC m=+1351.535838649" observedRunningTime="2026-03-13 14:27:34.749385828 +0000 UTC m=+1353.649173527" watchObservedRunningTime="2026-03-13 14:27:34.751821763 +0000 UTC m=+1353.651609462" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.774342 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.780920 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" podStartSLOduration=4.553303091 podStartE2EDuration="7.780899384s" podCreationTimestamp="2026-03-13 14:27:27 +0000 UTC" firstStartedPulling="2026-03-13 14:27:29.400143181 +0000 UTC m=+1348.299930870" lastFinishedPulling="2026-03-13 14:27:32.627739474 +0000 UTC m=+1351.527527163" observedRunningTime="2026-03-13 14:27:34.779247779 +0000 UTC m=+1353.679035478" watchObservedRunningTime="2026-03-13 14:27:34.780899384 +0000 UTC m=+1353.680687063" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.819131 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.885998 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:27:34 crc kubenswrapper[4907]: I0313 14:27:34.961826 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.686985 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerStarted","Data":"8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e"} Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.687263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerStarted","Data":"502cbc857cf878844cea8a77f9b3d530353b13301737ac4a045e3f57f1cad50e"} Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.712710 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.712997 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" containerID="cri-o://57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624" gracePeriod=30 Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.713056 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" containerID="cri-o://8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573" gracePeriod=30 Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.746718 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.747260 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.747502 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.747502 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.752577 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.756480 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.756808 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.767991 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.795936 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796023 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796271 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796318 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796366 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796485 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.796558 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.878353 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.878471 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.885706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899124 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899190 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899240 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899274 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899425 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.899456 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.908125 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.912596 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.913050 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.913331 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.915197 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.934621 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:35 crc kubenswrapper[4907]: I0313 14:27:35.936537 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"barbican-api-7fc86597fb-mt9pk\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.160599 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.724040 4907 generic.go:334] "Generic (PLEG): container finished" podID="070e2b75-05b1-42a4-a34e-950d45d65920" containerID="57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624" exitCode=143 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.724418 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerDied","Data":"57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624"} Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.737807 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" containerID="cri-o://6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.738941 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-c685c5475-rvkf2" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" containerID="cri-o://bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.739072 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerStarted","Data":"3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0"} Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.739459 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" containerID="cri-o://9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.739540 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-c685c5475-rvkf2" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" containerID="cri-o://2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4" gracePeriod=30 Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.777145 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bf587b8d9-pvvbz" podStartSLOduration=3.777128897 podStartE2EDuration="3.777128897s" podCreationTimestamp="2026-03-13 14:27:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:36.775219385 +0000 UTC m=+1355.675007074" watchObservedRunningTime="2026-03-13 14:27:36.777128897 +0000 UTC m=+1355.676916576" Mar 13 14:27:36 crc kubenswrapper[4907]: I0313 14:27:36.815202 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752502 4907 generic.go:334] "Generic (PLEG): container finished" podID="fcbb1687-db0b-4fc6-937c-89102831237b" containerID="9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5" exitCode=0 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752770 4907 generic.go:334] "Generic (PLEG): container finished" podID="fcbb1687-db0b-4fc6-937c-89102831237b" containerID="6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38" exitCode=143 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752592 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerDied","Data":"9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.752818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerDied","Data":"6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.754405 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerStarted","Data":"97dce4dda51605af3970178d1030468b7ba5970d94173d175f92a65d74507fff"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761159 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca4928ba-1277-4918-8b92-3eda2b276577" containerID="2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4" exitCode=0 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761193 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca4928ba-1277-4918-8b92-3eda2b276577" containerID="bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb" exitCode=143 Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerDied","Data":"2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.761293 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerDied","Data":"bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.769053 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerStarted","Data":"5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec"} Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.769325 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:27:37 crc kubenswrapper[4907]: I0313 14:27:37.805415 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7q999" podStartSLOduration=3.103559929 podStartE2EDuration="40.805393152s" podCreationTimestamp="2026-03-13 14:26:57 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.763220846 +0000 UTC m=+1317.663008535" lastFinishedPulling="2026-03-13 14:27:36.465054059 +0000 UTC m=+1355.364841758" observedRunningTime="2026-03-13 14:27:37.793297374 +0000 UTC m=+1356.693085063" watchObservedRunningTime="2026-03-13 14:27:37.805393152 +0000 UTC m=+1356.705180841" Mar 13 14:27:39 crc kubenswrapper[4907]: I0313 14:27:39.873057 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:39 crc kubenswrapper[4907]: I0313 14:27:39.959121 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:27:39 crc kubenswrapper[4907]: I0313 14:27:39.959680 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" containerID="cri-o://46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4" gracePeriod=10 Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.093693 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.258635 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.817544 4907 generic.go:334] "Generic (PLEG): container finished" podID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerID="46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4" exitCode=0 Mar 13 14:27:40 crc kubenswrapper[4907]: I0313 14:27:40.817608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerDied","Data":"46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4"} Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.153109 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:42076->10.217.0.160:9311: read: connection reset by peer" Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.153143 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:42066->10.217.0.160:9311: read: connection reset by peer" Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.755706 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.149:5353: connect: connection refused" Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.838244 4907 generic.go:334] "Generic (PLEG): container finished" podID="070e2b75-05b1-42a4-a34e-950d45d65920" containerID="8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573" exitCode=0 Mar 13 14:27:42 crc kubenswrapper[4907]: I0313 14:27:42.838289 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerDied","Data":"8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.044922 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.045148 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d49fd688d-njvfm" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": dial tcp 10.217.0.160:9311: connect: connection refused" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.867447 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" event={"ID":"fcbb1687-db0b-4fc6-937c-89102831237b","Type":"ContainerDied","Data":"7a66d01a94d06b4565675b1fcb3708a34133ebf672d54a557137e862ae4fd1c3"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.867749 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a66d01a94d06b4565675b1fcb3708a34133ebf672d54a557137e862ae4fd1c3" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.874430 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerStarted","Data":"f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.881626 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-c685c5475-rvkf2" event={"ID":"ca4928ba-1277-4918-8b92-3eda2b276577","Type":"ContainerDied","Data":"6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.881708 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c4a87798f44345ddbfe9a9bbadc067d17cd8c04974906e2bc3559decad063c2" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.884978 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d49fd688d-njvfm" event={"ID":"070e2b75-05b1-42a4-a34e-950d45d65920","Type":"ContainerDied","Data":"eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.885023 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaad4530bf1fa0795ee15397e0a9c4c262de8913b337803ac449643d1177e1b7" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.887282 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" event={"ID":"a661b9e1-c551-4b36-87a9-833df73bd6b2","Type":"ContainerDied","Data":"eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0"} Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.887319 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eaee566543e3557ec83d2d26f27876f1ccb4095ddf1c90e6680a5df6c9c175d0" Mar 13 14:27:43 crc kubenswrapper[4907]: I0313 14:27:43.953572 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.073121 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.073622 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074244 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074287 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074413 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.074527 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") pod \"a661b9e1-c551-4b36-87a9-833df73bd6b2\" (UID: \"a661b9e1-c551-4b36-87a9-833df73bd6b2\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.083194 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v" (OuterVolumeSpecName: "kube-api-access-ntd2v") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "kube-api-access-ntd2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.098833 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.132787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.139396 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.139720 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.141317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config" (OuterVolumeSpecName: "config") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.158668 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a661b9e1-c551-4b36-87a9-833df73bd6b2" (UID: "a661b9e1-c551-4b36-87a9-833df73bd6b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177409 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177468 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177525 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177552 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") pod \"070e2b75-05b1-42a4-a34e-950d45d65920\" (UID: \"070e2b75-05b1-42a4-a34e-950d45d65920\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.177994 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178014 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178025 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178035 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntd2v\" (UniqueName: \"kubernetes.io/projected/a661b9e1-c551-4b36-87a9-833df73bd6b2-kube-api-access-ntd2v\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178045 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178054 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a661b9e1-c551-4b36-87a9-833df73bd6b2-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.178099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs" (OuterVolumeSpecName: "logs") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.180311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.185068 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk" (OuterVolumeSpecName: "kube-api-access-bfwxk") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "kube-api-access-bfwxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.208742 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.222487 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.241986 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data" (OuterVolumeSpecName: "config-data") pod "070e2b75-05b1-42a4-a34e-950d45d65920" (UID: "070e2b75-05b1-42a4-a34e-950d45d65920"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.247211 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279254 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279288 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/070e2b75-05b1-42a4-a34e-950d45d65920-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279299 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfwxk\" (UniqueName: \"kubernetes.io/projected/070e2b75-05b1-42a4-a34e-950d45d65920-kube-api-access-bfwxk\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279308 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.279318 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/070e2b75-05b1-42a4-a34e-950d45d65920-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.380931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381363 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381507 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381911 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs" (OuterVolumeSpecName: "logs") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381918 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.381987 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") pod \"fcbb1687-db0b-4fc6-937c-89102831237b\" (UID: \"fcbb1687-db0b-4fc6-937c-89102831237b\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382094 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") pod \"ca4928ba-1277-4918-8b92-3eda2b276577\" (UID: \"ca4928ba-1277-4918-8b92-3eda2b276577\") " Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs" (OuterVolumeSpecName: "logs") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382580 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcbb1687-db0b-4fc6-937c-89102831237b-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.382605 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca4928ba-1277-4918-8b92-3eda2b276577-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.386405 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.386427 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx" (OuterVolumeSpecName: "kube-api-access-qttjx") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "kube-api-access-qttjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.388664 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.389112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx" (OuterVolumeSpecName: "kube-api-access-9vftx") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "kube-api-access-9vftx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.423416 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.430958 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.455682 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data" (OuterVolumeSpecName: "config-data") pod "fcbb1687-db0b-4fc6-937c-89102831237b" (UID: "fcbb1687-db0b-4fc6-937c-89102831237b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.459901 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data" (OuterVolumeSpecName: "config-data") pod "ca4928ba-1277-4918-8b92-3eda2b276577" (UID: "ca4928ba-1277-4918-8b92-3eda2b276577"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484526 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484582 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484593 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484603 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vftx\" (UniqueName: \"kubernetes.io/projected/ca4928ba-1277-4918-8b92-3eda2b276577-kube-api-access-9vftx\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484617 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qttjx\" (UniqueName: \"kubernetes.io/projected/fcbb1687-db0b-4fc6-937c-89102831237b-kube-api-access-qttjx\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484650 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcbb1687-db0b-4fc6-937c-89102831237b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484663 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.484674 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca4928ba-1277-4918-8b92-3eda2b276577-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.897282 4907 generic.go:334] "Generic (PLEG): container finished" podID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerID="5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec" exitCode=0 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.897369 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerDied","Data":"5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec"} Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.899765 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerStarted","Data":"22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07"} Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.899909 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.900014 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-54c69f4bdb-lt67r" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902512 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759cc7f497-7nv78" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902522 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-c685c5475-rvkf2" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902547 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerStarted","Data":"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1"} Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902779 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d49fd688d-njvfm" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902816 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" containerID="cri-o://b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902935 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" containerID="cri-o://af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902933 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" containerID="cri-o://dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902909 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" containerID="cri-o://a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" gracePeriod=30 Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.902963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.960752 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7fc86597fb-mt9pk" podStartSLOduration=9.96070637 podStartE2EDuration="9.96070637s" podCreationTimestamp="2026-03-13 14:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:44.950039259 +0000 UTC m=+1363.849826948" watchObservedRunningTime="2026-03-13 14:27:44.96070637 +0000 UTC m=+1363.860494069" Mar 13 14:27:44 crc kubenswrapper[4907]: I0313 14:27:44.981583 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.417911262 podStartE2EDuration="48.981562307s" podCreationTimestamp="2026-03-13 14:26:56 +0000 UTC" firstStartedPulling="2026-03-13 14:26:58.370075573 +0000 UTC m=+1317.269863262" lastFinishedPulling="2026-03-13 14:27:43.933726628 +0000 UTC m=+1362.833514307" observedRunningTime="2026-03-13 14:27:44.972362516 +0000 UTC m=+1363.872150225" watchObservedRunningTime="2026-03-13 14:27:44.981562307 +0000 UTC m=+1363.881349996" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.009220 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.021726 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-c685c5475-rvkf2"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.039116 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.050389 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-54c69f4bdb-lt67r"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.059821 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.068098 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-759cc7f497-7nv78"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.077333 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.086358 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5d49fd688d-njvfm"] Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.806716 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" path="/var/lib/kubelet/pods/070e2b75-05b1-42a4-a34e-950d45d65920/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.807330 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" path="/var/lib/kubelet/pods/a661b9e1-c551-4b36-87a9-833df73bd6b2/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.807945 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" path="/var/lib/kubelet/pods/ca4928ba-1277-4918-8b92-3eda2b276577/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.808988 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" path="/var/lib/kubelet/pods/fcbb1687-db0b-4fc6-937c-89102831237b/volumes" Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914169 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" exitCode=0 Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914223 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" exitCode=2 Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914235 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" exitCode=0 Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914258 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1"} Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914320 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b"} Mar 13 14:27:45 crc kubenswrapper[4907]: I0313 14:27:45.914337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c"} Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.254287 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.418613 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.418667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.418925 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419039 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419120 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419191 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419233 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") pod \"594b61fa-12b7-47b6-8af3-17f024e0d54d\" (UID: \"594b61fa-12b7-47b6-8af3-17f024e0d54d\") " Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.419713 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/594b61fa-12b7-47b6-8af3-17f024e0d54d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.425936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts" (OuterVolumeSpecName: "scripts") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.425980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.427063 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2" (OuterVolumeSpecName: "kube-api-access-wpft2") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "kube-api-access-wpft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.451700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.473125 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data" (OuterVolumeSpecName: "config-data") pod "594b61fa-12b7-47b6-8af3-17f024e0d54d" (UID: "594b61fa-12b7-47b6-8af3-17f024e0d54d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522589 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522664 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522682 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522697 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpft2\" (UniqueName: \"kubernetes.io/projected/594b61fa-12b7-47b6-8af3-17f024e0d54d-kube-api-access-wpft2\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.522712 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/594b61fa-12b7-47b6-8af3-17f024e0d54d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:46 crc kubenswrapper[4907]: I0313 14:27:46.872841 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.748925 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749097 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749131 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.749338 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") pod \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\" (UID: \"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5\") " Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.752645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.756717 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.781246 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw" (OuterVolumeSpecName: "kube-api-access-2xbmw") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "kube-api-access-2xbmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.781341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts" (OuterVolumeSpecName: "scripts") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888342 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888650 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xbmw\" (UniqueName: \"kubernetes.io/projected/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-kube-api-access-2xbmw\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.888668 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.907671 4907 generic.go:334] "Generic (PLEG): container finished" podID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" exitCode=0 Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.907805 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.927712 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.941691 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7q999" Mar 13 14:27:47 crc kubenswrapper[4907]: I0313 14:27:47.995239 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.006909 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512"} Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.006969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5","Type":"ContainerDied","Data":"cf9fd571ce09eea426c8836be80b3c4ef776ec41dabdcd7886212b3709f230ce"} Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.006986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7q999" event={"ID":"594b61fa-12b7-47b6-8af3-17f024e0d54d","Type":"ContainerDied","Data":"247e75fb8f70b076f7d10c60ec2902246989846bb77cff82137c6d0beb413598"} Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.007003 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="247e75fb8f70b076f7d10c60ec2902246989846bb77cff82137c6d0beb413598" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.007022 4907 scope.go:117] "RemoveContainer" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.034042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.046059 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.046122 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.075173 4907 scope.go:117] "RemoveContainer" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.100450 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.137265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data" (OuterVolumeSpecName: "config-data") pod "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" (UID: "58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.142959 4907 scope.go:117] "RemoveContainer" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.158650 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159080 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159098 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159113 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159120 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159132 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159142 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159150 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerName="cinder-db-sync" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159158 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerName="cinder-db-sync" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159168 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159174 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159188 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159195 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159209 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159216 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159230 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159237 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159251 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159259 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159269 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="init" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159275 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="init" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159300 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159307 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159326 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159333 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.159343 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159352 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159546 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159561 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcbb1687-db0b-4fc6-937c-89102831237b" containerName="barbican-keystone-listener-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159572 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a661b9e1-c551-4b36-87a9-833df73bd6b2" containerName="dnsmasq-dns" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159583 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" containerName="cinder-db-sync" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159601 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.159613 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api-log" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165009 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="sg-core" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165023 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-notification-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165038 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="070e2b75-05b1-42a4-a34e-950d45d65920" containerName="barbican-api" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165055 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca4928ba-1277-4918-8b92-3eda2b276577" containerName="barbican-worker" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165070 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="ceilometer-central-agent" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.165079 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" containerName="proxy-httpd" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.166157 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178165 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178483 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178608 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.178819 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d7phq" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.185758 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.201614 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.271687 4907 scope.go:117] "RemoveContainer" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303517 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303574 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303610 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.303655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.324968 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.335523 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.337256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.351415 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.366283 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.375156 4907 scope.go:117] "RemoveContainer" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.377058 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1\": container with ID starting with dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1 not found: ID does not exist" containerID="dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377101 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1"} err="failed to get container status \"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1\": rpc error: code = NotFound desc = could not find container \"dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1\": container with ID starting with dd889dbfbaa8a48ab7c7b9c0cf914f3824e4417de226256cac660ed3cfc140b1 not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377127 4907 scope.go:117] "RemoveContainer" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.377399 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b\": container with ID starting with a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b not found: ID does not exist" containerID="a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377417 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b"} err="failed to get container status \"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b\": rpc error: code = NotFound desc = could not find container \"a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b\": container with ID starting with a15bc8c355c09d50b27215622b3ea5aaebf1df184e4f387e2405fe44a886e49b not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.377428 4907 scope.go:117] "RemoveContainer" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.380133 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512\": container with ID starting with af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512 not found: ID does not exist" containerID="af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.380169 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512"} err="failed to get container status \"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512\": rpc error: code = NotFound desc = could not find container \"af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512\": container with ID starting with af56a2f9ef345a9cf4a63dafe33dfb02aa9518339a5d2ef4a11fbf9ec8ab7512 not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.380196 4907 scope.go:117] "RemoveContainer" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" Mar 13 14:27:48 crc kubenswrapper[4907]: E0313 14:27:48.391295 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c\": container with ID starting with b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c not found: ID does not exist" containerID="b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.391333 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c"} err="failed to get container status \"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c\": rpc error: code = NotFound desc = could not find container \"b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c\": container with ID starting with b9aab9a1290561a45c480e877db9222806553cdd6cfb12daf299ab965ad2ec5c not found: ID does not exist" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.391743 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.394241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.399789 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.400039 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411621 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411768 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.411947 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412089 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412173 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412199 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412221 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.412295 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.413899 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.414809 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.431735 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.432097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.432781 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.435786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.444011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"cinder-scheduler-0\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.466214 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.467979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.477330 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.507208 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517557 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517723 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517762 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.517794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.518916 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519104 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519137 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519173 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519275 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519300 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.519323 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.520234 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.520763 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.521343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.521288 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.541490 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"dnsmasq-dns-58b85ccffc-jt5lv\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.567812 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621129 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621214 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621254 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621316 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621333 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621365 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621389 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621411 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621432 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621447 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.621488 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.625579 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.630197 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.631097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.634413 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.634500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.634802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.638965 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"ceilometer-0\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.693360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723262 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723714 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723815 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723868 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.723911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.726188 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.726245 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.730101 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.732484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.736326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.736907 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.758075 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"cinder-api-0\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " pod="openstack/cinder-api-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.831583 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:27:48 crc kubenswrapper[4907]: I0313 14:27:48.890291 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.060566 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:49 crc kubenswrapper[4907]: W0313 14:27:49.072338 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7fd8554_798f_4e3a_acf1_8425c15e2665.slice/crio-58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5 WatchSource:0}: Error finding container 58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5: Status 404 returned error can't find the container with id 58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5 Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.262807 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:27:49 crc kubenswrapper[4907]: W0313 14:27:49.263684 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeed444b5_3ec0_44b6_b0a0_3e3e5a66fe24.slice/crio-a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33 WatchSource:0}: Error finding container a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33: Status 404 returned error can't find the container with id a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33 Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.439171 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.455824 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:27:49 crc kubenswrapper[4907]: I0313 14:27:49.808708 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5" path="/var/lib/kubelet/pods/58d1ad9f-02b2-43b0-92a3-8e28d8d1ddf5/volumes" Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.004402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"ad91253ab4bed9131d491de9a5560dd1867583453c957940978da876fc7db3cf"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.018094 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerStarted","Data":"bbe16c55906016b19d3449613b6670869e1b477f29a2b60cc4d61f1e4a53e85e"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.029319 4907 generic.go:334] "Generic (PLEG): container finished" podID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerID="4f693ef5bd7b7f8e96bb97b86b9be101cd4aab9df400b462e0a79f5818a5b73b" exitCode=0 Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.029385 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerDied","Data":"4f693ef5bd7b7f8e96bb97b86b9be101cd4aab9df400b462e0a79f5818a5b73b"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.029410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerStarted","Data":"a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33"} Mar 13 14:27:50 crc kubenswrapper[4907]: I0313 14:27:50.042750 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerStarted","Data":"58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.079763 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.082211 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.082237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.084313 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerStarted","Data":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.085781 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerStarted","Data":"6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.087051 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.097403 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerStarted","Data":"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d"} Mar 13 14:27:51 crc kubenswrapper[4907]: I0313 14:27:51.103694 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" podStartSLOduration=3.103677053 podStartE2EDuration="3.103677053s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:51.102543802 +0000 UTC m=+1370.002331491" watchObservedRunningTime="2026-03-13 14:27:51.103677053 +0000 UTC m=+1370.003464742" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerStarted","Data":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111711 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111390 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" containerID="cri-o://1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" gracePeriod=30 Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.111305 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" containerID="cri-o://e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" gracePeriod=30 Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.117035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerStarted","Data":"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59"} Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.120618 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a"} Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.140480 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.140462571 podStartE2EDuration="4.140462571s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:52.130586333 +0000 UTC m=+1371.030374022" watchObservedRunningTime="2026-03-13 14:27:52.140462571 +0000 UTC m=+1371.040250260" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.156712 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.570057367 podStartE2EDuration="4.156692862s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="2026-03-13 14:27:49.078811942 +0000 UTC m=+1367.978599641" lastFinishedPulling="2026-03-13 14:27:49.665447447 +0000 UTC m=+1368.565235136" observedRunningTime="2026-03-13 14:27:52.155261883 +0000 UTC m=+1371.055049582" watchObservedRunningTime="2026-03-13 14:27:52.156692862 +0000 UTC m=+1371.056480551" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.770668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.910807 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.910867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.910935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911083 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911432 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs" (OuterVolumeSpecName: "logs") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911835 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911779 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.911967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") pod \"ca2411de-0fb9-4702-aae0-808e5d759667\" (UID: \"ca2411de-0fb9-4702-aae0-808e5d759667\") " Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.912435 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca2411de-0fb9-4702-aae0-808e5d759667-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.912450 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ca2411de-0fb9-4702-aae0-808e5d759667-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.930317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk" (OuterVolumeSpecName: "kube-api-access-xlvfk") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "kube-api-access-xlvfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.931549 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts" (OuterVolumeSpecName: "scripts") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:52 crc kubenswrapper[4907]: I0313 14:27:52.952046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.019341 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.019574 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlvfk\" (UniqueName: \"kubernetes.io/projected/ca2411de-0fb9-4702-aae0-808e5d759667-kube-api-access-xlvfk\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.019667 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.033069 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.041139 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data" (OuterVolumeSpecName: "config-data") pod "ca2411de-0fb9-4702-aae0-808e5d759667" (UID: "ca2411de-0fb9-4702-aae0-808e5d759667"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.121792 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.121823 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca2411de-0fb9-4702-aae0-808e5d759667-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130355 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca2411de-0fb9-4702-aae0-808e5d759667" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" exitCode=0 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130386 4907 generic.go:334] "Generic (PLEG): container finished" podID="ca2411de-0fb9-4702-aae0-808e5d759667" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" exitCode=143 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130463 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerDied","Data":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130509 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerDied","Data":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ca2411de-0fb9-4702-aae0-808e5d759667","Type":"ContainerDied","Data":"bbe16c55906016b19d3449613b6670869e1b477f29a2b60cc4d61f1e4a53e85e"} Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.130608 4907 scope.go:117] "RemoveContainer" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.159712 4907 scope.go:117] "RemoveContainer" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.165191 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.183334 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.191556 4907 scope.go:117] "RemoveContainer" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.194129 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": container with ID starting with 1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d not found: ID does not exist" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194183 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} err="failed to get container status \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": rpc error: code = NotFound desc = could not find container \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": container with ID starting with 1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194216 4907 scope.go:117] "RemoveContainer" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.194675 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": container with ID starting with e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000 not found: ID does not exist" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194702 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} err="failed to get container status \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": rpc error: code = NotFound desc = could not find container \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": container with ID starting with e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000 not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194720 4907 scope.go:117] "RemoveContainer" containerID="1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d"} err="failed to get container status \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": rpc error: code = NotFound desc = could not find container \"1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d\": container with ID starting with 1a5d6bde83c26d0dd9c7eb5214a4d0bd35bb82277b346a9cbb3e5efef6ed9c8d not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.194985 4907 scope.go:117] "RemoveContainer" containerID="e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.195276 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000"} err="failed to get container status \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": rpc error: code = NotFound desc = could not find container \"e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000\": container with ID starting with e994224e3dd0dfe7ba5a1ff2805c3b640032239385a154735e1317449e6ca000 not found: ID does not exist" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.209716 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.210498 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210519 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" Mar 13 14:27:53 crc kubenswrapper[4907]: E0313 14:27:53.210545 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210555 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210799 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api-log" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.210831 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" containerName="cinder-api" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.212040 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.216005 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.217131 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.217365 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.226731 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227622 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227781 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.227974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.228125 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.228253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.233585 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.331365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.331471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.331551 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.332226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.332381 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.332584 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333240 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333313 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333364 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.333407 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.334403 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.334723 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.342500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.343248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.346369 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.347153 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.351398 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.353430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"cinder-api-0\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.408143 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.570379 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.583176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.649657 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.759466 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.759773 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74984bd8b4-x42pq" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" containerID="cri-o://3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b" gracePeriod=30 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.760439 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-74984bd8b4-x42pq" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" containerID="cri-o://be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250" gracePeriod=30 Mar 13 14:27:53 crc kubenswrapper[4907]: I0313 14:27:53.824097 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca2411de-0fb9-4702-aae0-808e5d759667" path="/var/lib/kubelet/pods/ca2411de-0fb9-4702-aae0-808e5d759667/volumes" Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.106462 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:27:54 crc kubenswrapper[4907]: W0313 14:27:54.110208 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41d7a0d1_a3ce_45f6_a8b7_fcf9de7f7a76.slice/crio-10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236 WatchSource:0}: Error finding container 10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236: Status 404 returned error can't find the container with id 10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236 Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.148846 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerStarted","Data":"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd"} Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.150206 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.151675 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerStarted","Data":"10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236"} Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.158999 4907 generic.go:334] "Generic (PLEG): container finished" podID="032183a7-c5db-4528-a211-803050007b68" containerID="3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b" exitCode=143 Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.159105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerDied","Data":"3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b"} Mar 13 14:27:54 crc kubenswrapper[4907]: I0313 14:27:54.178495 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.250185069 podStartE2EDuration="6.178475129s" podCreationTimestamp="2026-03-13 14:27:48 +0000 UTC" firstStartedPulling="2026-03-13 14:27:49.448725132 +0000 UTC m=+1368.348512821" lastFinishedPulling="2026-03-13 14:27:53.377015202 +0000 UTC m=+1372.276802881" observedRunningTime="2026-03-13 14:27:54.170162593 +0000 UTC m=+1373.069950282" watchObservedRunningTime="2026-03-13 14:27:54.178475129 +0000 UTC m=+1373.078262818" Mar 13 14:27:55 crc kubenswrapper[4907]: I0313 14:27:55.171375 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerStarted","Data":"2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae"} Mar 13 14:27:56 crc kubenswrapper[4907]: I0313 14:27:56.181995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerStarted","Data":"a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650"} Mar 13 14:27:56 crc kubenswrapper[4907]: I0313 14:27:56.214597 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.214577906 podStartE2EDuration="3.214577906s" podCreationTimestamp="2026-03-13 14:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:27:56.213935209 +0000 UTC m=+1375.113722898" watchObservedRunningTime="2026-03-13 14:27:56.214577906 +0000 UTC m=+1375.114365605" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.192515 4907 generic.go:334] "Generic (PLEG): container finished" podID="032183a7-c5db-4528-a211-803050007b68" containerID="be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250" exitCode=0 Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.193923 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerDied","Data":"be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250"} Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.193965 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.361412 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.532826 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533115 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533177 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533208 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533235 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") pod \"032183a7-c5db-4528-a211-803050007b68\" (UID: \"032183a7-c5db-4528-a211-803050007b68\") " Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.533874 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs" (OuterVolumeSpecName: "logs") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.534296 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/032183a7-c5db-4528-a211-803050007b68-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.540311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2" (OuterVolumeSpecName: "kube-api-access-tfsh2") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "kube-api-access-tfsh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.547168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.586722 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data" (OuterVolumeSpecName: "config-data") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.588347 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "032183a7-c5db-4528-a211-803050007b68" (UID: "032183a7-c5db-4528-a211-803050007b68"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635661 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635692 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfsh2\" (UniqueName: \"kubernetes.io/projected/032183a7-c5db-4528-a211-803050007b68-kube-api-access-tfsh2\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635704 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:57 crc kubenswrapper[4907]: I0313 14:27:57.635714 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/032183a7-c5db-4528-a211-803050007b68-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.203664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-74984bd8b4-x42pq" event={"ID":"032183a7-c5db-4528-a211-803050007b68","Type":"ContainerDied","Data":"3fa9f3f934d41bda9d2c0367bd05516b9fcfddd3de52bfe4494f0529db4eafd7"} Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.203713 4907 scope.go:117] "RemoveContainer" containerID="be905f5a0aa412e370260df6ece658648121619b19bad67aee3af44172b10250" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.203727 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-74984bd8b4-x42pq" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.229959 4907 scope.go:117] "RemoveContainer" containerID="3dc5784d660cfabfa43e3df490d1b2200ffa14fc3ff516c9488bcc67f04af55b" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.230775 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.240556 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-74984bd8b4-x42pq"] Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.697127 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.763068 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.763644 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" containerID="cri-o://10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f" gracePeriod=10 Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.939469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 14:27:58 crc kubenswrapper[4907]: I0313 14:27:58.998523 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.011675 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.076287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.218615 4907 generic.go:334] "Generic (PLEG): container finished" podID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerID="10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f" exitCode=0 Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.218702 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerDied","Data":"10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f"} Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.222038 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" containerID="cri-o://0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" gracePeriod=30 Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.222592 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" containerID="cri-o://2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" gracePeriod=30 Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.400491 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:27:59 crc kubenswrapper[4907]: E0313 14:27:59.401143 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401160 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" Mar 13 14:27:59 crc kubenswrapper[4907]: E0313 14:27:59.401173 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401179 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401327 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.401350 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="032183a7-c5db-4528-a211-803050007b68" containerName="barbican-api-log" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.402224 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.423757 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.428734 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479735 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479767 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479793 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479906 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.479931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.581254 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.581567 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.582607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.582721 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.582862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583090 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") pod \"49bc69d8-fe85-4994-a5cf-c51472a124e9\" (UID: \"49bc69d8-fe85-4994-a5cf-c51472a124e9\") " Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583364 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583486 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583807 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.583997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.584106 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.584404 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.584949 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.595434 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.599086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.599779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.601720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.612477 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.613650 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl" (OuterVolumeSpecName: "kube-api-access-r77nl") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "kube-api-access-r77nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.644699 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"placement-85dc486cc4-fwrm6\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.690085 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r77nl\" (UniqueName: \"kubernetes.io/projected/49bc69d8-fe85-4994-a5cf-c51472a124e9-kube-api-access-r77nl\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.739706 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.762721 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.766827 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.793106 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.793134 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.796674 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032183a7-c5db-4528-a211-803050007b68" path="/var/lib/kubelet/pods/032183a7-c5db-4528-a211-803050007b68/volumes" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.806336 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config" (OuterVolumeSpecName: "config") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.812754 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.823931 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "49bc69d8-fe85-4994-a5cf-c51472a124e9" (UID: "49bc69d8-fe85-4994-a5cf-c51472a124e9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.907528 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.907800 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:27:59 crc kubenswrapper[4907]: I0313 14:27:59.907812 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/49bc69d8-fe85-4994-a5cf-c51472a124e9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.050283 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.143954 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:28:00 crc kubenswrapper[4907]: E0313 14:28:00.144453 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="init" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.144482 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="init" Mar 13 14:28:00 crc kubenswrapper[4907]: E0313 14:28:00.144506 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.144516 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.144788 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" containerName="dnsmasq-dns" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.145622 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.149063 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.149074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.149433 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.151960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.257577 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.258362 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" event={"ID":"49bc69d8-fe85-4994-a5cf-c51472a124e9","Type":"ContainerDied","Data":"74e19d820a0370c065d553cb0dbb43ed632fe5214db6fe0a8a5f25c3d63f84ae"} Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.258416 4907 scope.go:117] "RemoveContainer" containerID="10817f472001c46a816f4b77aa0c7ef3e99533487d552c19bd292e7357506e9f" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.258427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc46d7df7-77dng" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.266737 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" exitCode=0 Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.266785 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerDied","Data":"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59"} Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.301631 4907 scope.go:117] "RemoveContainer" containerID="577048fe225fd9933f67497a9e39e7a61a8a9a2ce395fd1bd05afbdcf17490ca" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.304368 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:28:00 crc kubenswrapper[4907]: W0313 14:28:00.311356 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3ee069a_41fb_4cab_b650_9bb9c03ac271.slice/crio-80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710 WatchSource:0}: Error finding container 80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710: Status 404 returned error can't find the container with id 80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710 Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.313315 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.315583 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"auto-csr-approver-29556868-pzml6\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.325814 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fc46d7df7-77dng"] Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.417557 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"auto-csr-approver-29556868-pzml6\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.438440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"auto-csr-approver-29556868-pzml6\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.470108 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:00 crc kubenswrapper[4907]: I0313 14:28:00.913761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.280056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerStarted","Data":"baf4195c37e7df43009ff885852df0c99df75e43454c0fbdc4b11a230d7af5e3"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.282410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerStarted","Data":"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.282436 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerStarted","Data":"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.282451 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerStarted","Data":"80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710"} Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.283765 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.283797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.310076 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-85dc486cc4-fwrm6" podStartSLOduration=2.310056751 podStartE2EDuration="2.310056751s" podCreationTimestamp="2026-03-13 14:27:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:01.303869012 +0000 UTC m=+1380.203656701" watchObservedRunningTime="2026-03-13 14:28:01.310056751 +0000 UTC m=+1380.209844440" Mar 13 14:28:01 crc kubenswrapper[4907]: I0313 14:28:01.793505 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49bc69d8-fe85-4994-a5cf-c51472a124e9" path="/var/lib/kubelet/pods/49bc69d8-fe85-4994-a5cf-c51472a124e9/volumes" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.294749 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerStarted","Data":"52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d"} Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.313181 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556868-pzml6" podStartSLOduration=1.452517895 podStartE2EDuration="2.313161092s" podCreationTimestamp="2026-03-13 14:28:00 +0000 UTC" firstStartedPulling="2026-03-13 14:28:00.918018248 +0000 UTC m=+1379.817805937" lastFinishedPulling="2026-03-13 14:28:01.778661445 +0000 UTC m=+1380.678449134" observedRunningTime="2026-03-13 14:28:02.305619977 +0000 UTC m=+1381.205407666" watchObservedRunningTime="2026-03-13 14:28:02.313161092 +0000 UTC m=+1381.212948781" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.649011 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.651825 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.655606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.655733 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-c7qcz" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.655622 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.688086 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762107 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.762497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864478 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.864698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.866989 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.873150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.885412 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.889569 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"openstackclient\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.951964 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.952767 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.974557 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.987479 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.988952 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:02 crc kubenswrapper[4907]: I0313 14:28:02.997508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.104422 4907 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 14:28:03 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_87fb1bb3-8511-4d25-bf36-7ba0a278e101_0(90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616" Netns:"/var/run/netns/5e0c96e2-3668-4751-a7e4-6ac9d06b067f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616;K8S_POD_UID=87fb1bb3-8511-4d25-bf36-7ba0a278e101" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/87fb1bb3-8511-4d25-bf36-7ba0a278e101]: expected pod UID "87fb1bb3-8511-4d25-bf36-7ba0a278e101" but got "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" from Kube API Mar 13 14:28:03 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:28:03 crc kubenswrapper[4907]: > Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.104807 4907 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 14:28:03 crc kubenswrapper[4907]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_87fb1bb3-8511-4d25-bf36-7ba0a278e101_0(90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616" Netns:"/var/run/netns/5e0c96e2-3668-4751-a7e4-6ac9d06b067f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=90a3d91a8b93297c4e43bf3c51abc0a220d77d20297eda06a993ae96cc8b4616;K8S_POD_UID=87fb1bb3-8511-4d25-bf36-7ba0a278e101" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/87fb1bb3-8511-4d25-bf36-7ba0a278e101]: expected pod UID "87fb1bb3-8511-4d25-bf36-7ba0a278e101" but got "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" from Kube API Mar 13 14:28:03 crc kubenswrapper[4907]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 14:28:03 crc kubenswrapper[4907]: > pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.126803 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170297 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.170422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.271914 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272005 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272071 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272195 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272320 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") pod \"e7fd8554-798f-4e3a-acf1-8425c15e2665\" (UID: \"e7fd8554-798f-4e3a-acf1-8425c15e2665\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272619 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.272764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.274332 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.274441 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.274561 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.278392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.278714 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.280019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm" (OuterVolumeSpecName: "kube-api-access-r9xxm") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "kube-api-access-r9xxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.281122 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.283988 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts" (OuterVolumeSpecName: "scripts") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.284082 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e7fd8554-798f-4e3a-acf1-8425c15e2665-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.284763 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.307368 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"openstackclient\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.309693 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319020 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" exitCode=0 Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerDied","Data":"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d"} Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e7fd8554-798f-4e3a-acf1-8425c15e2665","Type":"ContainerDied","Data":"58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5"} Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319163 4907 scope.go:117] "RemoveContainer" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.319320 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.356153 4907 generic.go:334] "Generic (PLEG): container finished" podID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerID="52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d" exitCode=0 Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.356928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerDied","Data":"52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d"} Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.357023 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.380390 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.387467 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.387501 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9xxm\" (UniqueName: \"kubernetes.io/projected/e7fd8554-798f-4e3a-acf1-8425c15e2665-kube-api-access-r9xxm\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.387514 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.395494 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="87fb1bb3-8511-4d25-bf36-7ba0a278e101" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.397085 4907 scope.go:117] "RemoveContainer" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.404031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.448555 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data" (OuterVolumeSpecName: "config-data") pod "e7fd8554-798f-4e3a-acf1-8425c15e2665" (UID: "e7fd8554-798f-4e3a-acf1-8425c15e2665"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488396 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488541 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488604 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.488641 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") pod \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\" (UID: \"87fb1bb3-8511-4d25-bf36-7ba0a278e101\") " Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.489217 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.489238 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7fd8554-798f-4e3a-acf1-8425c15e2665-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.489499 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.493803 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts" (OuterVolumeSpecName: "kube-api-access-c9hts") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "kube-api-access-c9hts". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.495673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.495950 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87fb1bb3-8511-4d25-bf36-7ba0a278e101" (UID: "87fb1bb3-8511-4d25-bf36-7ba0a278e101"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.498574 4907 scope.go:117] "RemoveContainer" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.500896 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59\": container with ID starting with 2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59 not found: ID does not exist" containerID="2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.500937 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59"} err="failed to get container status \"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59\": rpc error: code = NotFound desc = could not find container \"2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59\": container with ID starting with 2b64ad381c9b4854fe824a41af04da817955826edfb608941d06803d7f8f4d59 not found: ID does not exist" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.500959 4907 scope.go:117] "RemoveContainer" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.502458 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d\": container with ID starting with 0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d not found: ID does not exist" containerID="0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.502486 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d"} err="failed to get container status \"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d\": rpc error: code = NotFound desc = could not find container \"0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d\": container with ID starting with 0a62f5f3cb31fd39481383f9dca4b2b20b3144ace647921ed05a9b4b9214a38d not found: ID does not exist" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592328 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592948 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9hts\" (UniqueName: \"kubernetes.io/projected/87fb1bb3-8511-4d25-bf36-7ba0a278e101-kube-api-access-c9hts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592974 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.592988 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87fb1bb3-8511-4d25-bf36-7ba0a278e101-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.658322 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.672218 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681123 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.681498 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681515 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.681522 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681528 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681730 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="cinder-scheduler" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.681753 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" containerName="probe" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.682949 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.685446 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.711117 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797373 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797511 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797578 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.797636 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.803050 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87fb1bb3-8511-4d25-bf36-7ba0a278e101" path="/var/lib/kubelet/pods/87fb1bb3-8511-4d25-bf36-7ba0a278e101/volumes" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.803510 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7fd8554-798f-4e3a-acf1-8425c15e2665" path="/var/lib/kubelet/pods/e7fd8554-798f-4e3a-acf1-8425c15e2665/volumes" Mar 13 14:28:03 crc kubenswrapper[4907]: E0313 14:28:03.888646 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7fd8554_798f_4e3a_acf1_8425c15e2665.slice/crio-58e8239ab4329277f0169cd2f4714646b1acec175437b4fe78554e7e213348e5\": RecentStats: unable to find data in memory cache]" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.898762 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.899895 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.901303 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.901442 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.901560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.902315 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.902459 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.902635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.906454 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.908117 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.908175 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.911756 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:03 crc kubenswrapper[4907]: I0313 14:28:03.922199 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"cinder-scheduler-0\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " pod="openstack/cinder-scheduler-0" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.012150 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.231204 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.320979 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.321359 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-944cbc5bb-fj29g" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" containerID="cri-o://a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75" gracePeriod=30 Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.321947 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-944cbc5bb-fj29g" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" containerID="cri-o://b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd" gracePeriod=30 Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.375390 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec","Type":"ContainerStarted","Data":"7886c1c9bb89142f6b353653e6db0031015d0af7793820ba26eb48966adcdb36"} Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.379285 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.399680 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="87fb1bb3-8511-4d25-bf36-7ba0a278e101" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.489518 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.750274 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.886327 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.895800 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556862-sczxs"] Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.923035 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") pod \"6301088c-1715-4d57-a9b8-1ea9f7128560\" (UID: \"6301088c-1715-4d57-a9b8-1ea9f7128560\") " Mar 13 14:28:04 crc kubenswrapper[4907]: I0313 14:28:04.928864 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg" (OuterVolumeSpecName: "kube-api-access-xmrsg") pod "6301088c-1715-4d57-a9b8-1ea9f7128560" (UID: "6301088c-1715-4d57-a9b8-1ea9f7128560"). InnerVolumeSpecName "kube-api-access-xmrsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.026129 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmrsg\" (UniqueName: \"kubernetes.io/projected/6301088c-1715-4d57-a9b8-1ea9f7128560-kube-api-access-xmrsg\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.399250 4907 generic.go:334] "Generic (PLEG): container finished" podID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerID="b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd" exitCode=0 Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.399625 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerDied","Data":"b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.403845 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556868-pzml6" event={"ID":"6301088c-1715-4d57-a9b8-1ea9f7128560","Type":"ContainerDied","Data":"baf4195c37e7df43009ff885852df0c99df75e43454c0fbdc4b11a230d7af5e3"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.403991 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baf4195c37e7df43009ff885852df0c99df75e43454c0fbdc4b11a230d7af5e3" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.404159 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556868-pzml6" Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.425061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerStarted","Data":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.425139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerStarted","Data":"8158569f0778590e4a7e4122294d9844f9402dad6a901826d7dd5af6e7fab659"} Mar 13 14:28:05 crc kubenswrapper[4907]: I0313 14:28:05.809014 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e27ed7-be4f-4d90-ac02-fa7d64585d6d" path="/var/lib/kubelet/pods/72e27ed7-be4f-4d90-ac02-fa7d64585d6d/volumes" Mar 13 14:28:06 crc kubenswrapper[4907]: I0313 14:28:06.108352 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 14:28:06 crc kubenswrapper[4907]: I0313 14:28:06.447585 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerStarted","Data":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} Mar 13 14:28:06 crc kubenswrapper[4907]: I0313 14:28:06.486176 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.486156596 podStartE2EDuration="3.486156596s" podCreationTimestamp="2026-03-13 14:28:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:06.481375467 +0000 UTC m=+1385.381163156" watchObservedRunningTime="2026-03-13 14:28:06.486156596 +0000 UTC m=+1385.385944295" Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.978244 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979061 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" containerID="cri-o://2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979255 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" containerID="cri-o://7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979300 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" containerID="cri-o://e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.979439 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" containerID="cri-o://041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" gracePeriod=30 Mar 13 14:28:07 crc kubenswrapper[4907]: I0313 14:28:07.995247 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.170:3000/\": EOF" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.340669 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:28:08 crc kubenswrapper[4907]: E0313 14:28:08.341677 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerName="oc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.344999 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerName="oc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.345452 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" containerName="oc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.346653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.349659 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.349847 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.351895 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.377215 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.471968 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" exitCode=0 Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472006 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" exitCode=2 Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472017 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" exitCode=0 Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472042 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd"} Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472090 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a"} Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.472105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e"} Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.507995 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508040 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508105 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508309 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508522 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.508666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.609966 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610066 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610102 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610132 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610194 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.610290 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.615469 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.616028 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.620831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.621316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.635075 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.640972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.642753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.645898 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"swift-proxy-7dfdbcdd5-dcb7n\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.672054 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:08 crc kubenswrapper[4907]: I0313 14:28:08.867728 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.013038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020259 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020373 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020412 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020640 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.020724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") pod \"f744eee5-064f-464c-86fc-5130071b313d\" (UID: \"f744eee5-064f-464c-86fc-5130071b313d\") " Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.024211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.024652 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.029851 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc" (OuterVolumeSpecName: "kube-api-access-dtpdc") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "kube-api-access-dtpdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.030287 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts" (OuterVolumeSpecName: "scripts") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.064022 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.116441 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128202 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtpdc\" (UniqueName: \"kubernetes.io/projected/f744eee5-064f-464c-86fc-5130071b313d-kube-api-access-dtpdc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128228 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128241 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128253 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128263 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f744eee5-064f-464c-86fc-5130071b313d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.128274 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.172171 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data" (OuterVolumeSpecName: "config-data") pod "f744eee5-064f-464c-86fc-5130071b313d" (UID: "f744eee5-064f-464c-86fc-5130071b313d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.229968 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f744eee5-064f-464c-86fc-5130071b313d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.363602 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.485368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerStarted","Data":"3472978398de27f6988019a31485f47afeca847038681d998902a9798bd82c2a"} Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488907 4907 generic.go:334] "Generic (PLEG): container finished" podID="f744eee5-064f-464c-86fc-5130071b313d" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" exitCode=0 Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5"} Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488976 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f744eee5-064f-464c-86fc-5130071b313d","Type":"ContainerDied","Data":"ad91253ab4bed9131d491de9a5560dd1867583453c957940978da876fc7db3cf"} Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.488991 4907 scope.go:117] "RemoveContainer" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.489120 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.538790 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.543217 4907 scope.go:117] "RemoveContainer" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.546653 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.568759 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569116 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569133 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569146 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569154 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569175 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.569197 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569203 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569354 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="proxy-httpd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569370 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="sg-core" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569389 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-central-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.569401 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f744eee5-064f-464c-86fc-5130071b313d" containerName="ceilometer-notification-agent" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.571138 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.577214 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.578444 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.584638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.591848 4907 scope.go:117] "RemoveContainer" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.623706 4907 scope.go:117] "RemoveContainer" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.652990 4907 scope.go:117] "RemoveContainer" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.656972 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd\": container with ID starting with 7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd not found: ID does not exist" containerID="7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657008 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd"} err="failed to get container status \"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd\": rpc error: code = NotFound desc = could not find container \"7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd\": container with ID starting with 7a11049e1abd6d70a403607531097f57571aa3ac1adb2b5af76d4fe1cd5fb6bd not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657029 4907 scope.go:117] "RemoveContainer" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.657424 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a\": container with ID starting with e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a not found: ID does not exist" containerID="e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657480 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a"} err="failed to get container status \"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a\": rpc error: code = NotFound desc = could not find container \"e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a\": container with ID starting with e47ab86336baef65bd0ea7edbd7c31a75a6d68df9c09a6a282295545b620225a not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.657547 4907 scope.go:117] "RemoveContainer" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.661014 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5\": container with ID starting with 041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5 not found: ID does not exist" containerID="041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.661070 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5"} err="failed to get container status \"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5\": rpc error: code = NotFound desc = could not find container \"041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5\": container with ID starting with 041e4dec51c360b9cb2fc26fee14b909f22ef44631885cdd0d5fc5f23a7ef2e5 not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.661099 4907 scope.go:117] "RemoveContainer" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" Mar 13 14:28:09 crc kubenswrapper[4907]: E0313 14:28:09.661418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e\": container with ID starting with 2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e not found: ID does not exist" containerID="2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.661447 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e"} err="failed to get container status \"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e\": rpc error: code = NotFound desc = could not find container \"2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e\": container with ID starting with 2ba377587d3943d9d1a24298ede2bf3a69a7589b2b93375a231da14a6aa5fa8e not found: ID does not exist" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739062 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739258 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739305 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.739324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.740547 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.740588 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.740621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.794758 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f744eee5-064f-464c-86fc-5130071b313d" path="/var/lib/kubelet/pods/f744eee5-064f-464c-86fc-5130071b313d/volumes" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.842837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.842992 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843028 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843101 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.843705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.849095 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.849688 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.852377 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.859652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.863753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"ceilometer-0\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " pod="openstack/ceilometer-0" Mar 13 14:28:09 crc kubenswrapper[4907]: I0313 14:28:09.898589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.497828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.504328 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerStarted","Data":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.504380 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerStarted","Data":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.506024 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.506139 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:10 crc kubenswrapper[4907]: I0313 14:28:10.532385 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" podStartSLOduration=2.5323603329999997 podStartE2EDuration="2.532360333s" podCreationTimestamp="2026-03-13 14:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:10.521713224 +0000 UTC m=+1389.421500913" watchObservedRunningTime="2026-03-13 14:28:10.532360333 +0000 UTC m=+1389.432148022" Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.518429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"2e74becf0a06fc2c8e64ca7acbdb2fa95c901f1325aa7aeacdbfe4e9db0c932e"} Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.731740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.732103 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" containerID="cri-o://6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5" gracePeriod=30 Mar 13 14:28:11 crc kubenswrapper[4907]: I0313 14:28:11.732415 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" containerID="cri-o://22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f" gracePeriod=30 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.505541 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.505796 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" containerID="cri-o://4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450" gracePeriod=30 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.506240 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" containerID="cri-o://456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f" gracePeriod=30 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.534910 4907 generic.go:334] "Generic (PLEG): container finished" podID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerID="6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5" exitCode=143 Mar 13 14:28:12 crc kubenswrapper[4907]: I0313 14:28:12.535024 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerDied","Data":"6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5"} Mar 13 14:28:13 crc kubenswrapper[4907]: I0313 14:28:13.561055 4907 generic.go:334] "Generic (PLEG): container finished" podID="5c36a352-151f-4e93-8094-3855bfed532e" containerID="4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450" exitCode=143 Mar 13 14:28:13 crc kubenswrapper[4907]: I0313 14:28:13.561134 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerDied","Data":"4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450"} Mar 13 14:28:13 crc kubenswrapper[4907]: I0313 14:28:13.721696 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:14 crc kubenswrapper[4907]: I0313 14:28:14.243159 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 14:28:14 crc kubenswrapper[4907]: I0313 14:28:14.897412 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:41636->10.217.0.154:9292: read: connection reset by peer" Mar 13 14:28:14 crc kubenswrapper[4907]: I0313 14:28:14.900155 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:41622->10.217.0.154:9292: read: connection reset by peer" Mar 13 14:28:15 crc kubenswrapper[4907]: I0313 14:28:15.591377 4907 generic.go:334] "Generic (PLEG): container finished" podID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerID="22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f" exitCode=0 Mar 13 14:28:15 crc kubenswrapper[4907]: I0313 14:28:15.591422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerDied","Data":"22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f"} Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.604044 4907 generic.go:334] "Generic (PLEG): container finished" podID="5c36a352-151f-4e93-8094-3855bfed532e" containerID="456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f" exitCode=0 Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.604096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerDied","Data":"456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f"} Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.615934 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.617031 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.654796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.722900 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.724510 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.730385 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.731671 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.734085 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.752354 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.769443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.795479 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.798700 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.798901 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.798989 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.799124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.799421 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.839755 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.842301 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.851238 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901321 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901486 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901638 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.901683 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.903074 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.906501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.908480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.920184 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.920301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.920360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.923710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.927933 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"nova-api-22bc-account-create-update-r969t\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.933765 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"nova-api-db-create-g96mx\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.941827 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"nova-cell0-db-create-vlp6r\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.943671 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.949667 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.960746 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.966243 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 14:28:16 crc kubenswrapper[4907]: I0313 14:28:16.975366 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022499 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.022575 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.023417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.044166 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"nova-cell1-db-create-ldbt9\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.056937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.065848 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.128848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.128908 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.129838 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.163091 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.166790 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"nova-cell0-f148-account-create-update-wgw4b\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.183476 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.185026 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.190366 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.193845 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.333415 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.335240 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.336178 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.435984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.436061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.436684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.458448 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"nova-cell1-f6fe-account-create-update-56786\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:17 crc kubenswrapper[4907]: I0313 14:28:17.520169 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.051146 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.051205 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.361999 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470565 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470759 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470830 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470859 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470941 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.470979 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.471014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") pod \"5c36a352-151f-4e93-8094-3855bfed532e\" (UID: \"5c36a352-151f-4e93-8094-3855bfed532e\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.472669 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs" (OuterVolumeSpecName: "logs") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.474131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.490567 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts" (OuterVolumeSpecName: "scripts") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.514601 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.524926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4" (OuterVolumeSpecName: "kube-api-access-f7gh4") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "kube-api-access-f7gh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575604 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575641 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575676 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575692 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5c36a352-151f-4e93-8094-3855bfed532e-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.575704 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7gh4\" (UniqueName: \"kubernetes.io/projected/5c36a352-151f-4e93-8094-3855bfed532e-kube-api-access-f7gh4\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.594944 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.630110 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.663723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec","Type":"ContainerStarted","Data":"3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.671129 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.684324 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.687496 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.687782 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.687923 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.702189 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.709656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5c36a352-151f-4e93-8094-3855bfed532e","Type":"ContainerDied","Data":"4cb91fefa1e0ea486010258682ba6df5f6a7f72f9a8eaedd21f297ae14bacf23"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.710135 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.710908 4907 scope.go:117] "RemoveContainer" containerID="456e65b2b22522c83f37e28d27a38f3af4c42d273550f8c8c1246de8d14f0a2f" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.711146 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.712923 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.761614914 podStartE2EDuration="16.712897554s" podCreationTimestamp="2026-03-13 14:28:02 +0000 UTC" firstStartedPulling="2026-03-13 14:28:03.903274809 +0000 UTC m=+1382.803062508" lastFinishedPulling="2026-03-13 14:28:17.854557459 +0000 UTC m=+1396.754345148" observedRunningTime="2026-03-13 14:28:18.685603681 +0000 UTC m=+1397.585391390" watchObservedRunningTime="2026-03-13 14:28:18.712897554 +0000 UTC m=+1397.612685273" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.717161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.736482 4907 generic.go:334] "Generic (PLEG): container finished" podID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerID="a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75" exitCode=0 Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.736573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerDied","Data":"a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75"} Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.785907 4907 scope.go:117] "RemoveContainer" containerID="4364cee9bed454e59183eb0004b9dfc4ab15dc5a345acbfd2a54032342758450" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.787606 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data" (OuterVolumeSpecName: "config-data") pod "5c36a352-151f-4e93-8094-3855bfed532e" (UID: "5c36a352-151f-4e93-8094-3855bfed532e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.797924 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.797973 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798486 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798546 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798783 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798818 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798873 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.798919 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") pod \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\" (UID: \"2c9261af-9eaa-4623-9f1b-719b58a9d3a2\") " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.799734 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c36a352-151f-4e93-8094-3855bfed532e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.802284 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs" (OuterVolumeSpecName: "logs") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.803035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.803199 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd" (OuterVolumeSpecName: "kube-api-access-7fdtd") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "kube-api-access-7fdtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.804026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.810490 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts" (OuterVolumeSpecName: "scripts") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.845050 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.865916 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.877026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data" (OuterVolumeSpecName: "config-data") pod "2c9261af-9eaa-4623-9f1b-719b58a9d3a2" (UID: "2c9261af-9eaa-4623-9f1b-719b58a9d3a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902112 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fdtd\" (UniqueName: \"kubernetes.io/projected/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-kube-api-access-7fdtd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902156 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902169 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902183 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902194 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902208 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902220 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c9261af-9eaa-4623-9f1b-719b58a9d3a2-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.902257 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 13 14:28:18 crc kubenswrapper[4907]: I0313 14:28:18.934705 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.005944 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.066936 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.115838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.135935 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136350 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136362 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136374 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136380 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136391 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136396 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.136421 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136427 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136593 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136604 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c36a352-151f-4e93-8094-3855bfed532e" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136617 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-log" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.136628 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" containerName="glance-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.137650 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.139166 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.141314 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.148937 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.175647 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210781 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210916 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.210955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211033 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") pod \"10a74a61-272c-4c54-9b2a-6379b77c9984\" (UID: \"10a74a61-272c-4c54-9b2a-6379b77c9984\") " Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211456 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211701 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211753 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211808 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211836 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211926 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.211957 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.212027 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.233005 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.235224 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt" (OuterVolumeSpecName: "kube-api-access-gsjwt") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "kube-api-access-gsjwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.245755 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.276573 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:28:19 crc kubenswrapper[4907]: W0313 14:28:19.276567 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4eeb2fb2_22fb_4ea8_a39b_5f959877cfe6.slice/crio-46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba WatchSource:0}: Error finding container 46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba: Status 404 returned error can't find the container with id 46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.292359 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.300597 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313202 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313249 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313321 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313361 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313426 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313443 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313513 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsjwt\" (UniqueName: \"kubernetes.io/projected/10a74a61-272c-4c54-9b2a-6379b77c9984-kube-api-access-gsjwt\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313523 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.313962 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.320210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.326465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.326728 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.328970 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:28:19 crc kubenswrapper[4907]: W0313 14:28:19.329155 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dcaf7c6_b4dd_4312_b68b_a833a827a2c2.slice/crio-fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819 WatchSource:0}: Error finding container fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819: Status 404 returned error can't find the container with id fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819 Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.331683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.333501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.337568 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.351622 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.389019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config" (OuterVolumeSpecName: "config") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.407201 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.414690 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.445645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.462549 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.472076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "10a74a61-272c-4c54-9b2a-6379b77c9984" (UID: "10a74a61-272c-4c54-9b2a-6379b77c9984"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.491685 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.516519 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.517104 4907 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a74a61-272c-4c54-9b2a-6379b77c9984-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.749982 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-22bc-account-create-update-r969t" event={"ID":"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6","Type":"ContainerStarted","Data":"46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.762508 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2c9261af-9eaa-4623-9f1b-719b58a9d3a2","Type":"ContainerDied","Data":"2cde1712ac814253467a7ebeac79ad28f1b6bc8f06f7979a4876a959f01886bd"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.762562 4907 scope.go:117] "RemoveContainer" containerID="22707601406e7d79a6ae3140c68d6ec6d23c0168f7d4a1523dfe2e817903318f" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.762723 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.843213 4907 scope.go:117] "RemoveContainer" containerID="6af4b6388e126ccd4e3d2cc04b6747e5283033aaa5d424f9e2907953b3f55ea5" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.845369 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c36a352-151f-4e93-8094-3855bfed532e" path="/var/lib/kubelet/pods/5c36a352-151f-4e93-8094-3855bfed532e/volumes" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.851922 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-944cbc5bb-fj29g" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871203 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871253 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871275 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871292 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerStarted","Data":"fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871315 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerStarted","Data":"3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerStarted","Data":"b75ab2628fbe2653fbc966d7e517d3692b8792f5300df76a97698a45b1204d93"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871343 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-944cbc5bb-fj29g" event={"ID":"10a74a61-272c-4c54-9b2a-6379b77c9984","Type":"ContainerDied","Data":"b6a3399278c9114cdb41d239ecf9266ca04769841425ab359d3ec15a0ea2369a"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871358 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerStarted","Data":"107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.871371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerStarted","Data":"57e7cdfdc101c4e36f2572edec180e1f998c526cf20ca796f61ebb46ef192472"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.875696 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-ldbt9" podStartSLOduration=3.875669518 podStartE2EDuration="3.875669518s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.845110048 +0000 UTC m=+1398.744897727" watchObservedRunningTime="2026-03-13 14:28:19.875669518 +0000 UTC m=+1398.775457207" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.890357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerStarted","Data":"13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.890402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerStarted","Data":"5c53d7426e0e1950f78a3c70586fa0b2e353687a87ba5dbe4ff6e6fc12256f18"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.920168 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerStarted","Data":"ca4cf56d486af5cc064a60fd28eb3806cfa9fe3d0bef16b65c3fd811e37a4cb4"} Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.943637 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.944662 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.944681 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" Mar 13 14:28:19 crc kubenswrapper[4907]: E0313 14:28:19.944713 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.944723 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.945107 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-api" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.945141 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" containerName="neutron-httpd" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.946787 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.949408 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.964834 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.977746 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:19 crc kubenswrapper[4907]: I0313 14:28:19.991132 4907 scope.go:117] "RemoveContainer" containerID="b624bd9d00cbcadf5cdc9520721c5131708177fac0d6006945f6f3ffc27df6cd" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.023587 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-vlp6r" podStartSLOduration=4.02353542 podStartE2EDuration="4.02353542s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.883156582 +0000 UTC m=+1398.782944271" watchObservedRunningTime="2026-03-13 14:28:20.02353542 +0000 UTC m=+1398.923323119" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.060001 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-g96mx" podStartSLOduration=4.059974091 podStartE2EDuration="4.059974091s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.904661177 +0000 UTC m=+1398.804448876" watchObservedRunningTime="2026-03-13 14:28:20.059974091 +0000 UTC m=+1398.959761780" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.072070 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-f6fe-account-create-update-56786" podStartSLOduration=3.072045449 podStartE2EDuration="3.072045449s" podCreationTimestamp="2026-03-13 14:28:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.93933101 +0000 UTC m=+1398.839118699" watchObservedRunningTime="2026-03-13 14:28:20.072045449 +0000 UTC m=+1398.971833138" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.103134 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.109222 4907 scope.go:117] "RemoveContainer" containerID="a7d63138378c0ce32dc24735fd0ab7db352755405ba476da24cb2065858f3b75" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130467 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130514 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130548 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130569 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130616 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130637 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130679 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.130752 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.136534 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-944cbc5bb-fj29g"] Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.162391 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" podStartSLOduration=4.162371505 podStartE2EDuration="4.162371505s" podCreationTimestamp="2026-03-13 14:28:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:19.983522881 +0000 UTC m=+1398.883310580" watchObservedRunningTime="2026-03-13 14:28:20.162371505 +0000 UTC m=+1399.062159194" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.203352 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232344 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232479 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232546 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232619 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.232681 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.233153 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.234673 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.234783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.238483 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.238625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.238693 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.241543 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.256728 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.285505 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.347815 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.931765 4907 generic.go:334] "Generic (PLEG): container finished" podID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerID="e156e90b4d1cbea6f3a55f14df88e09d8aeafa1d0449cf8de460be879c418cf0" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.932918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-22bc-account-create-update-r969t" event={"ID":"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6","Type":"ContainerDied","Data":"e156e90b4d1cbea6f3a55f14df88e09d8aeafa1d0449cf8de460be879c418cf0"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.934262 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerStarted","Data":"d3f2a65d466df300e932d28e3a4a5fb0f68b272bc80ae37edb5a080aa7144cae"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.937406 4907 generic.go:334] "Generic (PLEG): container finished" podID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerID="107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.937589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerDied","Data":"107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.939402 4907 generic.go:334] "Generic (PLEG): container finished" podID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerID="13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.939474 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerDied","Data":"13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.944233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerStarted","Data":"598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.954694 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerStarted","Data":"ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188"} Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.958203 4907 generic.go:334] "Generic (PLEG): container finished" podID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerID="3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3" exitCode=0 Mar 13 14:28:20 crc kubenswrapper[4907]: I0313 14:28:20.958268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerDied","Data":"3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.012686 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.139845 4907 scope.go:117] "RemoveContainer" containerID="ff874f38ed077cec3c5651ce4ba0b1fcf1996e7c127b7d6b128f6c5552e26bed" Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.822951 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a74a61-272c-4c54-9b2a-6379b77c9984" path="/var/lib/kubelet/pods/10a74a61-272c-4c54-9b2a-6379b77c9984/volumes" Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.823980 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c9261af-9eaa-4623-9f1b-719b58a9d3a2" path="/var/lib/kubelet/pods/2c9261af-9eaa-4623-9f1b-719b58a9d3a2/volumes" Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.973311 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerStarted","Data":"eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.973361 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerStarted","Data":"0ede32ea598e7a7a739c54b190ba597842e8cec1fea7546f52eefbb2b1fa0f83"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.976721 4907 generic.go:334] "Generic (PLEG): container finished" podID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerID="ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188" exitCode=0 Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.976793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerDied","Data":"ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.985987 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.989235 4907 generic.go:334] "Generic (PLEG): container finished" podID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerID="598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035" exitCode=0 Mar 13 14:28:21 crc kubenswrapper[4907]: I0313 14:28:21.989798 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerDied","Data":"598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035"} Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.008177 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerStarted","Data":"70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8"} Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.590525 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.691744 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") pod \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.692142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") pod \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\" (UID: \"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.693349 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" (UID: "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.698992 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m" (OuterVolumeSpecName: "kube-api-access-4pm5m") pod "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" (UID: "4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6"). InnerVolumeSpecName "kube-api-access-4pm5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.795248 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pm5m\" (UniqueName: \"kubernetes.io/projected/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-kube-api-access-4pm5m\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.795280 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.806126 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.832719 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.834280 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.896594 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") pod \"7af00fde-8a92-4530-8fa8-9df7ff787c93\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.896789 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") pod \"7af00fde-8a92-4530-8fa8-9df7ff787c93\" (UID: \"7af00fde-8a92-4530-8fa8-9df7ff787c93\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.897299 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7af00fde-8a92-4530-8fa8-9df7ff787c93" (UID: "7af00fde-8a92-4530-8fa8-9df7ff787c93"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.902261 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc" (OuterVolumeSpecName: "kube-api-access-dn4dc") pod "7af00fde-8a92-4530-8fa8-9df7ff787c93" (UID: "7af00fde-8a92-4530-8fa8-9df7ff787c93"). InnerVolumeSpecName "kube-api-access-dn4dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998038 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") pod \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998400 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") pod \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\" (UID: \"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998568 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") pod \"251ab76e-b9df-406f-9b07-c7fa6e227feb\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998609 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") pod \"251ab76e-b9df-406f-9b07-c7fa6e227feb\" (UID: \"251ab76e-b9df-406f-9b07-c7fa6e227feb\") " Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.998976 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7af00fde-8a92-4530-8fa8-9df7ff787c93-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.999018 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn4dc\" (UniqueName: \"kubernetes.io/projected/7af00fde-8a92-4530-8fa8-9df7ff787c93-kube-api-access-dn4dc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.999177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" (UID: "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:22 crc kubenswrapper[4907]: I0313 14:28:22.999466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "251ab76e-b9df-406f-9b07-c7fa6e227feb" (UID: "251ab76e-b9df-406f-9b07-c7fa6e227feb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.002705 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k" (OuterVolumeSpecName: "kube-api-access-swn2k") pod "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" (UID: "9118baf0-df6a-41e5-98e5-3e4d8edfbaaa"). InnerVolumeSpecName "kube-api-access-swn2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.006593 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8" (OuterVolumeSpecName: "kube-api-access-mjhh8") pod "251ab76e-b9df-406f-9b07-c7fa6e227feb" (UID: "251ab76e-b9df-406f-9b07-c7fa6e227feb"). InnerVolumeSpecName "kube-api-access-mjhh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.033682 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f6fe-account-create-update-56786" event={"ID":"7af00fde-8a92-4530-8fa8-9df7ff787c93","Type":"ContainerDied","Data":"5c53d7426e0e1950f78a3c70586fa0b2e353687a87ba5dbe4ff6e6fc12256f18"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.033718 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c53d7426e0e1950f78a3c70586fa0b2e353687a87ba5dbe4ff6e6fc12256f18" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.033775 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f6fe-account-create-update-56786" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.068853 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vlp6r" event={"ID":"9118baf0-df6a-41e5-98e5-3e4d8edfbaaa","Type":"ContainerDied","Data":"b75ab2628fbe2653fbc966d7e517d3692b8792f5300df76a97698a45b1204d93"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.068925 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b75ab2628fbe2653fbc966d7e517d3692b8792f5300df76a97698a45b1204d93" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.068993 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vlp6r" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.096566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerStarted","Data":"0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100621 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100653 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/251ab76e-b9df-406f-9b07-c7fa6e227feb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100668 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjhh8\" (UniqueName: \"kubernetes.io/projected/251ab76e-b9df-406f-9b07-c7fa6e227feb-kube-api-access-mjhh8\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.100709 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swn2k\" (UniqueName: \"kubernetes.io/projected/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa-kube-api-access-swn2k\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.115214 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-22bc-account-create-update-r969t" event={"ID":"4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6","Type":"ContainerDied","Data":"46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.115477 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46336f79de33964f45a2c37528d8472d8d5d46341b5f9fc4bc15531e76f002ba" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.115620 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-22bc-account-create-update-r969t" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.137794 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.137776329 podStartE2EDuration="4.137776329s" podCreationTimestamp="2026-03-13 14:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:23.133082051 +0000 UTC m=+1402.032869750" watchObservedRunningTime="2026-03-13 14:28:23.137776329 +0000 UTC m=+1402.037564018" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.141660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerStarted","Data":"479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.152259 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g96mx" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.152518 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g96mx" event={"ID":"251ab76e-b9df-406f-9b07-c7fa6e227feb","Type":"ContainerDied","Data":"57e7cdfdc101c4e36f2572edec180e1f998c526cf20ca796f61ebb46ef192472"} Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.152596 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57e7cdfdc101c4e36f2572edec180e1f998c526cf20ca796f61ebb46ef192472" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.192428 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.192410305 podStartE2EDuration="4.192410305s" podCreationTimestamp="2026-03-13 14:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:23.187486351 +0000 UTC m=+1402.087274040" watchObservedRunningTime="2026-03-13 14:28:23.192410305 +0000 UTC m=+1402.092197994" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.610374 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.713536 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.720038 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") pod \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.720446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") pod \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\" (UID: \"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.721154 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" (UID: "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.721744 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.724545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn" (OuterVolumeSpecName: "kube-api-access-wj4tn") pod "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" (UID: "5dcaf7c6-b4dd-4312-b68b-a833a827a2c2"). InnerVolumeSpecName "kube-api-access-wj4tn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.823528 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") pod \"85913727-bbcd-4343-9faf-a75f40b42dc8\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.823684 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") pod \"85913727-bbcd-4343-9faf-a75f40b42dc8\" (UID: \"85913727-bbcd-4343-9faf-a75f40b42dc8\") " Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.824079 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wj4tn\" (UniqueName: \"kubernetes.io/projected/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2-kube-api-access-wj4tn\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.824119 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "85913727-bbcd-4343-9faf-a75f40b42dc8" (UID: "85913727-bbcd-4343-9faf-a75f40b42dc8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.827684 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw" (OuterVolumeSpecName: "kube-api-access-cnlpw") pod "85913727-bbcd-4343-9faf-a75f40b42dc8" (UID: "85913727-bbcd-4343-9faf-a75f40b42dc8"). InnerVolumeSpecName "kube-api-access-cnlpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.926046 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnlpw\" (UniqueName: \"kubernetes.io/projected/85913727-bbcd-4343-9faf-a75f40b42dc8-kube-api-access-cnlpw\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:23 crc kubenswrapper[4907]: I0313 14:28:23.926086 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/85913727-bbcd-4343-9faf-a75f40b42dc8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.171157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ldbt9" event={"ID":"5dcaf7c6-b4dd-4312-b68b-a833a827a2c2","Type":"ContainerDied","Data":"fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819"} Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.171199 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ldbt9" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.171227 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa6acba78f266f6ff1774ad62b03c8a5d55736cb5f7697c6fcc3a1a5a0614819" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerStarted","Data":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175721 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" containerID="cri-o://53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175841 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" containerID="cri-o://725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175919 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" containerID="cri-o://7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.175958 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" containerID="cri-o://306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" gracePeriod=30 Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.176192 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.181212 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.182224 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f148-account-create-update-wgw4b" event={"ID":"85913727-bbcd-4343-9faf-a75f40b42dc8","Type":"ContainerDied","Data":"ca4cf56d486af5cc064a60fd28eb3806cfa9fe3d0bef16b65c3fd811e37a4cb4"} Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.182292 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca4cf56d486af5cc064a60fd28eb3806cfa9fe3d0bef16b65c3fd811e37a4cb4" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.211208 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.461854472 podStartE2EDuration="15.211189153s" podCreationTimestamp="2026-03-13 14:28:09 +0000 UTC" firstStartedPulling="2026-03-13 14:28:10.510785786 +0000 UTC m=+1389.410573475" lastFinishedPulling="2026-03-13 14:28:23.260120467 +0000 UTC m=+1402.159908156" observedRunningTime="2026-03-13 14:28:24.200235215 +0000 UTC m=+1403.100022904" watchObservedRunningTime="2026-03-13 14:28:24.211189153 +0000 UTC m=+1403.110976842" Mar 13 14:28:24 crc kubenswrapper[4907]: E0313 14:28:24.446388 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7ca16f6_c9d9_4415_b8cf_b4753d784170.slice/crio-725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20.scope\": RecentStats: unable to find data in memory cache]" Mar 13 14:28:24 crc kubenswrapper[4907]: I0313 14:28:24.931015 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.049904 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050238 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050327 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050398 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050545 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.050676 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") pod \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\" (UID: \"f7ca16f6-c9d9-4415-b8cf-b4753d784170\") " Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051051 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051444 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.051483 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f7ca16f6-c9d9-4415-b8cf-b4753d784170-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.056937 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts" (OuterVolumeSpecName: "scripts") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.065636 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm" (OuterVolumeSpecName: "kube-api-access-gfnvm") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "kube-api-access-gfnvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.085649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.127849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154022 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154056 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154068 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.154079 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfnvm\" (UniqueName: \"kubernetes.io/projected/f7ca16f6-c9d9-4415-b8cf-b4753d784170-kube-api-access-gfnvm\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.156622 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data" (OuterVolumeSpecName: "config-data") pod "f7ca16f6-c9d9-4415-b8cf-b4753d784170" (UID: "f7ca16f6-c9d9-4415-b8cf-b4753d784170"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198306 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" exitCode=0 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198338 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" exitCode=2 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198351 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" exitCode=0 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198360 4907 generic.go:334] "Generic (PLEG): container finished" podID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" exitCode=0 Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198384 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198431 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198439 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f7ca16f6-c9d9-4415-b8cf-b4753d784170","Type":"ContainerDied","Data":"2e74becf0a06fc2c8e64ca7acbdb2fa95c901f1325aa7aeacdbfe4e9db0c932e"} Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.198452 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.199042 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.255661 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7ca16f6-c9d9-4415-b8cf-b4753d784170-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.259948 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.261027 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.294567 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.307631 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308052 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308070 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308083 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308090 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308101 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308109 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308121 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308128 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308145 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308150 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308162 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308167 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308182 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308188 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308201 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308208 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308217 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308223 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.308236 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308242 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308398 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308410 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-central-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308422 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308435 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" containerName="mariadb-account-create-update" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308442 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="ceilometer-notification-agent" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308452 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308460 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="sg-core" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308480 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" containerName="mariadb-database-create" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.308488 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" containerName="proxy-httpd" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.310733 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.316579 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.317102 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.317424 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.323415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.353820 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358837 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358911 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.358975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.359013 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.359080 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.359104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.383818 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.384287 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384313 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384337 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.384636 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384671 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.384696 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.385055 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385138 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385212 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: E0313 14:28:25.385515 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385594 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385663 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.385919 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386001 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386239 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386314 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.386958 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387068 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387443 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387659 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.387735 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388090 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388204 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388451 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388530 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388744 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.388825 4907 scope.go:117] "RemoveContainer" containerID="725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389050 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20"} err="failed to get container status \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": rpc error: code = NotFound desc = could not find container \"725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20\": container with ID starting with 725fda3e17763bb514d9ba2d39dd7a4cfc02d5d585475411f3905020ac1f8e20 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389136 4907 scope.go:117] "RemoveContainer" containerID="7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389348 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e"} err="failed to get container status \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": rpc error: code = NotFound desc = could not find container \"7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e\": container with ID starting with 7ca97160ecf2400b113e2721d2509a72b44110e201fc18a9c3d14fa335b64d1e not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389423 4907 scope.go:117] "RemoveContainer" containerID="306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389640 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6"} err="failed to get container status \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": rpc error: code = NotFound desc = could not find container \"306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6\": container with ID starting with 306d706e8b95e9a38af5ba674a473f59462492d56df0ecb87578b8edfdde9aa6 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389715 4907 scope.go:117] "RemoveContainer" containerID="53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.389965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2"} err="failed to get container status \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": rpc error: code = NotFound desc = could not find container \"53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2\": container with ID starting with 53d201dd158c6af9aadd5213934f6cc439d1b5e00b0a3e3af8f7f6824b4594d2 not found: ID does not exist" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461233 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461467 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461605 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461660 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461697 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461726 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.461757 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.462093 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.465847 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.466347 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.466858 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.469720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.478484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"ceilometer-0\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.654064 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:25 crc kubenswrapper[4907]: I0313 14:28:25.832804 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7ca16f6-c9d9-4415-b8cf-b4753d784170" path="/var/lib/kubelet/pods/f7ca16f6-c9d9-4415-b8cf-b4753d784170/volumes" Mar 13 14:28:26 crc kubenswrapper[4907]: I0313 14:28:26.157559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:26 crc kubenswrapper[4907]: W0313 14:28:26.161848 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6be28d7_de73_47ad_9b71_f7333bae7487.slice/crio-1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a WatchSource:0}: Error finding container 1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a: Status 404 returned error can't find the container with id 1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a Mar 13 14:28:26 crc kubenswrapper[4907]: I0313 14:28:26.208835 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a"} Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.221461 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21"} Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.237230 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.238594 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.240162 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r4dpz" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.240500 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.241677 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.249785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.320993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.321295 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.321360 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.321529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.423853 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.423981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.424138 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.424171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.432204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.443292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.451704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.453483 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"nova-cell0-conductor-db-sync-z92h5\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.557717 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:27 crc kubenswrapper[4907]: I0313 14:28:27.838349 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:28 crc kubenswrapper[4907]: I0313 14:28:28.076290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:28:28 crc kubenswrapper[4907]: W0313 14:28:28.079392 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7ea7a2_ab25_4cb4_8749_10bc129cbd22.slice/crio-7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463 WatchSource:0}: Error finding container 7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463: Status 404 returned error can't find the container with id 7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463 Mar 13 14:28:28 crc kubenswrapper[4907]: I0313 14:28:28.247226 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerStarted","Data":"7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463"} Mar 13 14:28:28 crc kubenswrapper[4907]: I0313 14:28:28.248913 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81"} Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.265863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e"} Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.493183 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.493239 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.546524 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:29 crc kubenswrapper[4907]: I0313 14:28:29.553654 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282312 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerStarted","Data":"c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99"} Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282863 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282896 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282895 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" containerID="cri-o://af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282928 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" containerID="cri-o://7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282584 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" containerID="cri-o://d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.282811 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" containerID="cri-o://c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99" gracePeriod=30 Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.307752 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.768228928 podStartE2EDuration="5.307731263s" podCreationTimestamp="2026-03-13 14:28:25 +0000 UTC" firstStartedPulling="2026-03-13 14:28:26.164107428 +0000 UTC m=+1405.063895117" lastFinishedPulling="2026-03-13 14:28:29.703609753 +0000 UTC m=+1408.603397452" observedRunningTime="2026-03-13 14:28:30.303623792 +0000 UTC m=+1409.203411481" watchObservedRunningTime="2026-03-13 14:28:30.307731263 +0000 UTC m=+1409.207518952" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.348720 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.348779 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.381702 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:28:30 crc kubenswrapper[4907]: I0313 14:28:30.404239 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.090458 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.293716 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99" exitCode=0 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.294150 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e" exitCode=2 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.294170 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81" exitCode=0 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.293946 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99"} Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295258 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e"} Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295277 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81"} Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295777 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.295938 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.324959 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.415934 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.416376 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-67d6ccd9c4-ms98l" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" containerID="cri-o://e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6" gracePeriod=30 Mar 13 14:28:31 crc kubenswrapper[4907]: I0313 14:28:31.416695 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-67d6ccd9c4-ms98l" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" containerID="cri-o://38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2" gracePeriod=30 Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.305974 4907 generic.go:334] "Generic (PLEG): container finished" podID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerID="e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6" exitCode=143 Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.306094 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerDied","Data":"e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6"} Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.708782 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:32 crc kubenswrapper[4907]: I0313 14:28:32.709246 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.036537 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.314479 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.314505 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.847513 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:28:33 crc kubenswrapper[4907]: I0313 14:28:33.849974 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 14:28:34 crc kubenswrapper[4907]: E0313 14:28:34.716720 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d14b2b_1d0e_4e15_b837_0d2fb6aed785.slice/crio-conmon-38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2.scope\": RecentStats: unable to find data in memory cache]" Mar 13 14:28:35 crc kubenswrapper[4907]: I0313 14:28:35.338957 4907 generic.go:334] "Generic (PLEG): container finished" podID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerID="38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2" exitCode=0 Mar 13 14:28:35 crc kubenswrapper[4907]: I0313 14:28:35.340188 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerDied","Data":"38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2"} Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.400523 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.410604 4907 generic.go:334] "Generic (PLEG): container finished" podID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerID="d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21" exitCode=0 Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.410662 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21"} Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.413158 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67d6ccd9c4-ms98l" event={"ID":"45d14b2b-1d0e-4e15-b837-0d2fb6aed785","Type":"ContainerDied","Data":"0497ac4d8503d6e45cbe04af5eff7d939b60b6a6232917e037fb11fdbb11b8dd"} Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.413183 4907 scope.go:117] "RemoveContainer" containerID="38b798d627fc6c621c04314f998e3ab08f943e11ebed179dd2fc6641c6c999a2" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.413314 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67d6ccd9c4-ms98l" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441726 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441805 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441918 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441960 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.441986 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.442034 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") pod \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\" (UID: \"45d14b2b-1d0e-4e15-b837-0d2fb6aed785\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.448672 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs" (OuterVolumeSpecName: "logs") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.454025 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts" (OuterVolumeSpecName: "scripts") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.454449 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr" (OuterVolumeSpecName: "kube-api-access-4sstr") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "kube-api-access-4sstr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.466454 4907 scope.go:117] "RemoveContainer" containerID="e2ad3bad04a843173cf63dc1e67628c441f9aeb3514e5f9bf8afda3eba7b52f6" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.508269 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.513140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data" (OuterVolumeSpecName: "config-data") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543744 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543773 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543783 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sstr\" (UniqueName: \"kubernetes.io/projected/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-kube-api-access-4sstr\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543791 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.543800 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.571996 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.597123 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.616314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "45d14b2b-1d0e-4e15-b837-0d2fb6aed785" (UID: "45d14b2b-1d0e-4e15-b837-0d2fb6aed785"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644398 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644454 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644512 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644549 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644936 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.644956 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45d14b2b-1d0e-4e15-b837-0d2fb6aed785-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.645368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.649966 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc" (OuterVolumeSpecName: "kube-api-access-fldgc") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "kube-api-access-fldgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.651786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts" (OuterVolumeSpecName: "scripts") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.744213 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data" (OuterVolumeSpecName: "config-data") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.745996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746038 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746060 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") pod \"d6be28d7-de73-47ad-9b71-f7333bae7487\" (UID: \"d6be28d7-de73-47ad-9b71-f7333bae7487\") " Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746497 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746516 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fldgc\" (UniqueName: \"kubernetes.io/projected/d6be28d7-de73-47ad-9b71-f7333bae7487-kube-api-access-fldgc\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746526 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746536 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.746669 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.749770 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.760189 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-67d6ccd9c4-ms98l"] Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.768582 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.812548 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6be28d7-de73-47ad-9b71-f7333bae7487" (UID: "d6be28d7-de73-47ad-9b71-f7333bae7487"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.848689 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.848749 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6be28d7-de73-47ad-9b71-f7333bae7487-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:38 crc kubenswrapper[4907]: I0313 14:28:38.848760 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6be28d7-de73-47ad-9b71-f7333bae7487-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.426020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6be28d7-de73-47ad-9b71-f7333bae7487","Type":"ContainerDied","Data":"1554f3d915bdc2d1f39f5b0de39037d95cef463ea1345bb0b4172dd3d3d6666a"} Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.426096 4907 scope.go:117] "RemoveContainer" containerID="c34a9a5b2a3bb1f13c00c35e532469604057426d78bfb43c38cf1c8fd94a7c99" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.426111 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.431203 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerStarted","Data":"d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682"} Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.460736 4907 scope.go:117] "RemoveContainer" containerID="7174b12084675db6df525f79c405a1d67339b8435a26c05853001bd9082cd33e" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.463866 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-z92h5" podStartSLOduration=2.129845158 podStartE2EDuration="12.463839136s" podCreationTimestamp="2026-03-13 14:28:27 +0000 UTC" firstStartedPulling="2026-03-13 14:28:28.081499466 +0000 UTC m=+1406.981287155" lastFinishedPulling="2026-03-13 14:28:38.415493444 +0000 UTC m=+1417.315281133" observedRunningTime="2026-03-13 14:28:39.456911378 +0000 UTC m=+1418.356699067" watchObservedRunningTime="2026-03-13 14:28:39.463839136 +0000 UTC m=+1418.363626815" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.481147 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.487777 4907 scope.go:117] "RemoveContainer" containerID="af061011b60951ed9fad2d1d58dc7db5393d0bd0c5f54d1158edbedf15496e81" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.498754 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505136 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505511 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505527 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505555 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505562 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505574 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505581 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505589 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505595 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505616 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505621 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" Mar 13 14:28:39 crc kubenswrapper[4907]: E0313 14:28:39.505633 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505638 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505791 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="proxy-httpd" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505803 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-api" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505814 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-notification-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505826 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="sg-core" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505836 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" containerName="ceilometer-central-agent" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.505846 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" containerName="placement-log" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.507502 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.516938 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.519276 4907 scope.go:117] "RemoveContainer" containerID="d6e624a00a594aeaa75ca4ca30dbf27af1166a2309f4a98440260e91beadde21" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.519334 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.519372 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561454 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561476 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.561524 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663122 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663144 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663286 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.663985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.664065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.669493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.670164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.670898 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.678627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.681446 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"ceilometer-0\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " pod="openstack/ceilometer-0" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.799044 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45d14b2b-1d0e-4e15-b837-0d2fb6aed785" path="/var/lib/kubelet/pods/45d14b2b-1d0e-4e15-b837-0d2fb6aed785/volumes" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.799651 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6be28d7-de73-47ad-9b71-f7333bae7487" path="/var/lib/kubelet/pods/d6be28d7-de73-47ad-9b71-f7333bae7487/volumes" Mar 13 14:28:39 crc kubenswrapper[4907]: I0313 14:28:39.843643 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:40 crc kubenswrapper[4907]: W0313 14:28:40.286160 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8e62ba8_78aa_4acd_92f3_5adf43ca85ad.slice/crio-a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7 WatchSource:0}: Error finding container a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7: Status 404 returned error can't find the container with id a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7 Mar 13 14:28:40 crc kubenswrapper[4907]: I0313 14:28:40.292917 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:40 crc kubenswrapper[4907]: I0313 14:28:40.445451 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7"} Mar 13 14:28:41 crc kubenswrapper[4907]: I0313 14:28:41.454808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} Mar 13 14:28:42 crc kubenswrapper[4907]: I0313 14:28:42.467704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} Mar 13 14:28:43 crc kubenswrapper[4907]: I0313 14:28:43.477392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} Mar 13 14:28:45 crc kubenswrapper[4907]: I0313 14:28:45.495292 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerStarted","Data":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} Mar 13 14:28:45 crc kubenswrapper[4907]: I0313 14:28:45.496966 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:28:45 crc kubenswrapper[4907]: I0313 14:28:45.522347 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.485405578 podStartE2EDuration="6.522328702s" podCreationTimestamp="2026-03-13 14:28:39 +0000 UTC" firstStartedPulling="2026-03-13 14:28:40.289101731 +0000 UTC m=+1419.188889420" lastFinishedPulling="2026-03-13 14:28:44.326024855 +0000 UTC m=+1423.225812544" observedRunningTime="2026-03-13 14:28:45.521364465 +0000 UTC m=+1424.421152154" watchObservedRunningTime="2026-03-13 14:28:45.522328702 +0000 UTC m=+1424.422116391" Mar 13 14:28:47 crc kubenswrapper[4907]: I0313 14:28:47.207609 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.041153 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.041212 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.041263 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.042101 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.042164 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57" gracePeriod=600 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526004 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57" exitCode=0 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526705 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" containerID="cri-o://59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" gracePeriod=30 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526069 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57"} Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.526807 4907 scope.go:117] "RemoveContainer" containerID="ac423129eeb1fa8eb8df325acb5ccb91e119992cd551dd9d8c11d304cb95b736" Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.527199 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" containerID="cri-o://6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" gracePeriod=30 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.527240 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" containerID="cri-o://4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" gracePeriod=30 Mar 13 14:28:48 crc kubenswrapper[4907]: I0313 14:28:48.527272 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" containerID="cri-o://f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" gracePeriod=30 Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543246 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" exitCode=0 Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543767 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" exitCode=2 Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} Mar 13 14:28:49 crc kubenswrapper[4907]: I0313 14:28:49.543839 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.258234 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.317782 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318026 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318117 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318171 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.318405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") pod \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\" (UID: \"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad\") " Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.320926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.321262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.326262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7" (OuterVolumeSpecName: "kube-api-access-hdhb7") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "kube-api-access-hdhb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.326846 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts" (OuterVolumeSpecName: "scripts") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.353553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.418383 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421523 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421577 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421590 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421604 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421615 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdhb7\" (UniqueName: \"kubernetes.io/projected/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-kube-api-access-hdhb7\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.421624 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.470146 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data" (OuterVolumeSpecName: "config-data") pod "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" (UID: "b8e62ba8-78aa-4acd-92f3-5adf43ca85ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.523670 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553852 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" exitCode=0 Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553892 4907 generic.go:334] "Generic (PLEG): container finished" podID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" exitCode=0 Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553948 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.553990 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.554002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8e62ba8-78aa-4acd-92f3-5adf43ca85ad","Type":"ContainerDied","Data":"a83f85b89f55541dd31712e7fe8b295e79d7ff0a8e724ec9e202e87b0b0f97d7"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.554018 4907 scope.go:117] "RemoveContainer" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.558169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2"} Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.587115 4907 scope.go:117] "RemoveContainer" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.612303 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.620390 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.620686 4907 scope.go:117] "RemoveContainer" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637008 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637687 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637706 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637720 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637726 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637745 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637752 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.637765 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637772 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637958 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="proxy-httpd" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637976 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-notification-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.637988 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="sg-core" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.638007 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" containerName="ceilometer-central-agent" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.639745 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.644451 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.644662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.655681 4907 scope.go:117] "RemoveContainer" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.655698 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.691439 4907 scope.go:117] "RemoveContainer" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.691764 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": container with ID starting with 6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082 not found: ID does not exist" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.691794 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} err="failed to get container status \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": rpc error: code = NotFound desc = could not find container \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": container with ID starting with 6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.691815 4907 scope.go:117] "RemoveContainer" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.692289 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": container with ID starting with 4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf not found: ID does not exist" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692309 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} err="failed to get container status \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": rpc error: code = NotFound desc = could not find container \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": container with ID starting with 4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692339 4907 scope.go:117] "RemoveContainer" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.692691 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": container with ID starting with f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763 not found: ID does not exist" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692713 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} err="failed to get container status \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": rpc error: code = NotFound desc = could not find container \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": container with ID starting with f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.692736 4907 scope.go:117] "RemoveContainer" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: E0313 14:28:50.693177 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": container with ID starting with 59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723 not found: ID does not exist" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693198 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} err="failed to get container status \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": rpc error: code = NotFound desc = could not find container \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": container with ID starting with 59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693209 4907 scope.go:117] "RemoveContainer" containerID="6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693503 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082"} err="failed to get container status \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": rpc error: code = NotFound desc = could not find container \"6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082\": container with ID starting with 6ece6a9813874bfabeea08b5871d5ce4883ed931c7898fc7cf4565120db9c082 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693518 4907 scope.go:117] "RemoveContainer" containerID="4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693745 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf"} err="failed to get container status \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": rpc error: code = NotFound desc = could not find container \"4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf\": container with ID starting with 4b6c6996dbdabf0608354c555365a022e97acac7a151f533aa94e6c52bcba1bf not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.693765 4907 scope.go:117] "RemoveContainer" containerID="f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.697117 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763"} err="failed to get container status \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": rpc error: code = NotFound desc = could not find container \"f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763\": container with ID starting with f069ee7a6257ccd2ec10dd948c5d3a36fd6dadfff2902dc16abe2e27eff98763 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.697139 4907 scope.go:117] "RemoveContainer" containerID="59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.697675 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723"} err="failed to get container status \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": rpc error: code = NotFound desc = could not find container \"59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723\": container with ID starting with 59220582376a55be5a4d3d891ace1fbaa90206be7f65a48006e97d333f02f723 not found: ID does not exist" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.726734 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.726773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727020 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727099 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727198 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.727319 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.828970 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829072 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829738 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.829997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.831192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.831309 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.836984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.838028 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.841501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.847694 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.850076 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"ceilometer-0\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " pod="openstack/ceilometer-0" Mar 13 14:28:50 crc kubenswrapper[4907]: I0313 14:28:50.958424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.395413 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:28:51 crc kubenswrapper[4907]: W0313 14:28:51.400484 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda85017ea_b984_471a_a9a8_22c02d94448b.slice/crio-5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6 WatchSource:0}: Error finding container 5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6: Status 404 returned error can't find the container with id 5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6 Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.570702 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6"} Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.573700 4907 generic.go:334] "Generic (PLEG): container finished" podID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerID="d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682" exitCode=0 Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.573796 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerDied","Data":"d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682"} Mar 13 14:28:51 crc kubenswrapper[4907]: I0313 14:28:51.793074 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8e62ba8-78aa-4acd-92f3-5adf43ca85ad" path="/var/lib/kubelet/pods/b8e62ba8-78aa-4acd-92f3-5adf43ca85ad/volumes" Mar 13 14:28:52 crc kubenswrapper[4907]: I0313 14:28:52.586851 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a"} Mar 13 14:28:52 crc kubenswrapper[4907]: I0313 14:28:52.964939 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.125840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.125913 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.126172 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.126209 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") pod \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\" (UID: \"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22\") " Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.143075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv" (OuterVolumeSpecName: "kube-api-access-dc5bv") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "kube-api-access-dc5bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.147237 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts" (OuterVolumeSpecName: "scripts") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.197281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.210268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data" (OuterVolumeSpecName: "config-data") pod "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" (UID: "4d7ea7a2-ab25-4cb4-8749-10bc129cbd22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228234 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc5bv\" (UniqueName: \"kubernetes.io/projected/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-kube-api-access-dc5bv\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228270 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228279 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.228288 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.598034 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7"} Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.598396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01"} Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.600035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z92h5" event={"ID":"4d7ea7a2-ab25-4cb4-8749-10bc129cbd22","Type":"ContainerDied","Data":"7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463"} Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.600059 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d0624634945e5c5e075e91b43232f99c08cdffbe02fc1416ac047abb7f58463" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.600109 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z92h5" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.716468 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:28:53 crc kubenswrapper[4907]: E0313 14:28:53.717237 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerName="nova-cell0-conductor-db-sync" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.717282 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerName="nova-cell0-conductor-db-sync" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.717670 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" containerName="nova-cell0-conductor-db-sync" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.718741 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.721420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-r4dpz" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.721697 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.735147 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.842134 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.842223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.842315 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.943319 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.943398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.943457 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.948211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.948239 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:53 crc kubenswrapper[4907]: I0313 14:28:53.960509 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"nova-cell0-conductor-0\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:54 crc kubenswrapper[4907]: I0313 14:28:54.051109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:54 crc kubenswrapper[4907]: I0313 14:28:54.529180 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:28:54 crc kubenswrapper[4907]: I0313 14:28:54.611937 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerStarted","Data":"416b8c99953bb29e1da48539894292de008485b785183dda880949342468afcc"} Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.624858 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerStarted","Data":"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f"} Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.625396 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.629514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerStarted","Data":"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9"} Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.629776 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.643970 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.6439525120000003 podStartE2EDuration="2.643952512s" podCreationTimestamp="2026-03-13 14:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:28:55.64095317 +0000 UTC m=+1434.540740859" watchObservedRunningTime="2026-03-13 14:28:55.643952512 +0000 UTC m=+1434.543740211" Mar 13 14:28:55 crc kubenswrapper[4907]: I0313 14:28:55.668932 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.14756681 podStartE2EDuration="5.668887839s" podCreationTimestamp="2026-03-13 14:28:50 +0000 UTC" firstStartedPulling="2026-03-13 14:28:51.403261757 +0000 UTC m=+1430.303049446" lastFinishedPulling="2026-03-13 14:28:54.924582786 +0000 UTC m=+1433.824370475" observedRunningTime="2026-03-13 14:28:55.664474679 +0000 UTC m=+1434.564262368" watchObservedRunningTime="2026-03-13 14:28:55.668887839 +0000 UTC m=+1434.568675528" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.080126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.543417 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.544690 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.547063 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.547650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.553242 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.656746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.657123 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.657152 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.657180 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762098 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762188 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762213 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.762241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.770823 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.770899 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.774667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.779288 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.780384 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"nova-cell0-cell-mapping-6g25t\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.780432 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.783181 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.832869 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.864802 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.894220 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.902181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.914741 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.930180 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.954941 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.956426 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.963395 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.964809 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.964912 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.965058 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:28:59 crc kubenswrapper[4907]: I0313 14:28:59.984769 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.068704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069283 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069401 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069756 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069837 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.069937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.070052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.070143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.074610 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.076322 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.097232 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.097758 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.097987 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.134925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"nova-scheduler-0\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.150259 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175193 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175216 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175244 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175314 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175407 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175436 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.175825 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.192678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.192700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.194250 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.207447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.207875 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.213183 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279376 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279489 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279518 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.279607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.280900 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"nova-metadata-0\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.281168 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.293029 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.294473 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.303915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.304168 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.305720 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.321782 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.333463 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"nova-api-0\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.366342 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381063 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381123 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381151 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381549 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.381660 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.479245 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483273 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483314 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483343 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.483366 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.484240 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.484614 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.485037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.485181 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.485570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.503060 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"dnsmasq-dns-69b4446475-ttj9k\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.609037 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.647381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.683788 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerStarted","Data":"63f23d24127f696c6ac8a975f8b88fae3bcf7cd62b9d4f19bb8d47d4b8611d9a"} Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.866073 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.936425 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.938381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.944514 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.944769 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 14:29:00 crc kubenswrapper[4907]: I0313 14:29:00.953502 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.005444 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:01 crc kubenswrapper[4907]: W0313 14:29:01.008114 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1915a375_806d_43bb_8749_9524db071714.slice/crio-878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8 WatchSource:0}: Error finding container 878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8: Status 404 returned error can't find the container with id 878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8 Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.106933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.107004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.107033 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.107056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.209840 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.212757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.212791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.212857 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.222232 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.224495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.225462 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.227966 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.235749 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.241643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"nova-cell1-conductor-db-sync-xw7n9\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.270705 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.424684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:01 crc kubenswrapper[4907]: W0313 14:29:01.448149 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf56f7387_a1c9_4250_8abf_6c0a3830970e.slice/crio-c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5 WatchSource:0}: Error finding container c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5: Status 404 returned error can't find the container with id c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5 Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.712819 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerStarted","Data":"f5827c9400a454954bce2169d1bdd16e49d4ec2488819960057446785871a720"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.718863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerStarted","Data":"c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.723170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerStarted","Data":"878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.724923 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerStarted","Data":"bc36702cf525d044a0ffd3094a7d069c598da5cdc98d37c88eb61db1ff303142"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.732790 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerStarted","Data":"62294b7ca355e8d04aaa630081ec7abb20dd7d9bb3d9f97d4610e23fae479816"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.738973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerStarted","Data":"dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4"} Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.831607 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-6g25t" podStartSLOduration=2.8315857810000002 podStartE2EDuration="2.831585781s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:01.816097301 +0000 UTC m=+1440.715885010" watchObservedRunningTime="2026-03-13 14:29:01.831585781 +0000 UTC m=+1440.731373470" Mar 13 14:29:01 crc kubenswrapper[4907]: I0313 14:29:01.868576 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.766078 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerStarted","Data":"28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c"} Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.767231 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerStarted","Data":"dc2a57efb2cfb970e90ab542c28a0ad4000fce19809d4298c8e7c3ed553951bb"} Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.770960 4907 generic.go:334] "Generic (PLEG): container finished" podID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" exitCode=0 Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.771027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerDied","Data":"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624"} Mar 13 14:29:02 crc kubenswrapper[4907]: I0313 14:29:02.786570 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" podStartSLOduration=2.786537005 podStartE2EDuration="2.786537005s" podCreationTimestamp="2026-03-13 14:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:02.78268486 +0000 UTC m=+1441.682472549" watchObservedRunningTime="2026-03-13 14:29:02.786537005 +0000 UTC m=+1441.686324694" Mar 13 14:29:04 crc kubenswrapper[4907]: I0313 14:29:04.468362 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:04 crc kubenswrapper[4907]: I0313 14:29:04.540643 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.812104 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerStarted","Data":"0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.814568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerStarted","Data":"7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.814615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerStarted","Data":"7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817790 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerStarted","Data":"ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817823 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" containerID="cri-o://c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac" gracePeriod=30 Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817837 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerStarted","Data":"c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.817942 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" containerID="cri-o://ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961" gracePeriod=30 Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.829018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerStarted","Data":"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.829383 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.831356 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.929577322 podStartE2EDuration="6.831345623s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:00.88447607 +0000 UTC m=+1439.784263759" lastFinishedPulling="2026-03-13 14:29:04.786244371 +0000 UTC m=+1443.686032060" observedRunningTime="2026-03-13 14:29:05.831089606 +0000 UTC m=+1444.730877315" watchObservedRunningTime="2026-03-13 14:29:05.831345623 +0000 UTC m=+1444.731133312" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.836656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerStarted","Data":"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285"} Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.836766 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" gracePeriod=30 Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.859596 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" podStartSLOduration=5.8595774800000004 podStartE2EDuration="5.85957748s" podCreationTimestamp="2026-03-13 14:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:05.856748843 +0000 UTC m=+1444.756536532" watchObservedRunningTime="2026-03-13 14:29:05.85957748 +0000 UTC m=+1444.759365169" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.876678 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.337270115 podStartE2EDuration="6.876656364s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:01.249176095 +0000 UTC m=+1440.148963784" lastFinishedPulling="2026-03-13 14:29:04.788562354 +0000 UTC m=+1443.688350033" observedRunningTime="2026-03-13 14:29:05.874145625 +0000 UTC m=+1444.773933314" watchObservedRunningTime="2026-03-13 14:29:05.876656364 +0000 UTC m=+1444.776444053" Mar 13 14:29:05 crc kubenswrapper[4907]: I0313 14:29:05.921300 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.376429297 podStartE2EDuration="6.921280076s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:01.241386252 +0000 UTC m=+1440.141173941" lastFinishedPulling="2026-03-13 14:29:04.786237031 +0000 UTC m=+1443.686024720" observedRunningTime="2026-03-13 14:29:05.888580787 +0000 UTC m=+1444.788368476" watchObservedRunningTime="2026-03-13 14:29:05.921280076 +0000 UTC m=+1444.821067765" Mar 13 14:29:06 crc kubenswrapper[4907]: I0313 14:29:06.849283 4907 generic.go:334] "Generic (PLEG): container finished" podID="a99d655e-f6db-4c83-a950-b935588b7df1" containerID="c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac" exitCode=143 Mar 13 14:29:06 crc kubenswrapper[4907]: I0313 14:29:06.849998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerDied","Data":"c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac"} Mar 13 14:29:09 crc kubenswrapper[4907]: I0313 14:29:09.879855 4907 generic.go:334] "Generic (PLEG): container finished" podID="1b701a42-f649-4657-9250-49188b10d3b9" containerID="dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4" exitCode=0 Mar 13 14:29:09 crc kubenswrapper[4907]: I0313 14:29:09.879944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerDied","Data":"dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4"} Mar 13 14:29:09 crc kubenswrapper[4907]: I0313 14:29:09.911042 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=7.133873359 podStartE2EDuration="10.911026676s" podCreationTimestamp="2026-03-13 14:28:59 +0000 UTC" firstStartedPulling="2026-03-13 14:29:01.009811644 +0000 UTC m=+1439.909599323" lastFinishedPulling="2026-03-13 14:29:04.786964951 +0000 UTC m=+1443.686752640" observedRunningTime="2026-03-13 14:29:05.956229415 +0000 UTC m=+1444.856017104" watchObservedRunningTime="2026-03-13 14:29:09.911026676 +0000 UTC m=+1448.810814365" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.208827 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.208892 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.237324 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.322922 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.480872 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.480943 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.649252 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.710062 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.710296 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" containerID="cri-o://6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63" gracePeriod=10 Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.894866 4907 generic.go:334] "Generic (PLEG): container finished" podID="65556420-dc7f-49c8-af59-abafebbb4674" containerID="28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c" exitCode=0 Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.894964 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerDied","Data":"28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c"} Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.898967 4907 generic.go:334] "Generic (PLEG): container finished" podID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerID="6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63" exitCode=0 Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.900157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerDied","Data":"6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63"} Mar 13 14:29:10 crc kubenswrapper[4907]: I0313 14:29:10.942175 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.474709 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.480743 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556103 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556305 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556347 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556450 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") pod \"1b701a42-f649-4657-9250-49188b10d3b9\" (UID: \"1b701a42-f649-4657-9250-49188b10d3b9\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556490 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556515 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.556594 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") pod \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\" (UID: \"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24\") " Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.566047 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.566180 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.575567 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts" (OuterVolumeSpecName: "scripts") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.575693 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh" (OuterVolumeSpecName: "kube-api-access-5zvgh") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "kube-api-access-5zvgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.582356 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn" (OuterVolumeSpecName: "kube-api-access-jrsdn") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "kube-api-access-jrsdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.604437 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.640146 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data" (OuterVolumeSpecName: "config-data") pod "1b701a42-f649-4657-9250-49188b10d3b9" (UID: "1b701a42-f649-4657-9250-49188b10d3b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.641545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.656429 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config" (OuterVolumeSpecName: "config") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.657695 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.658734 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.661923 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zvgh\" (UniqueName: \"kubernetes.io/projected/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-kube-api-access-5zvgh\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662153 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662391 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662682 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662762 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b701a42-f649-4657-9250-49188b10d3b9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.662833 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.663270 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.663340 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.663422 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrsdn\" (UniqueName: \"kubernetes.io/projected/1b701a42-f649-4657-9250-49188b10d3b9-kube-api-access-jrsdn\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.673436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" (UID: "eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.765444 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.910994 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.910987 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b85ccffc-jt5lv" event={"ID":"eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24","Type":"ContainerDied","Data":"a66bc3530699af62851033529fd1a73503f7565c8f14e919806b488f53d8de33"} Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.911596 4907 scope.go:117] "RemoveContainer" containerID="6b06c4d709787d5a6f17229ea98351d8fef051feb5e0a2046eb23fd42751cf63" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.914547 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-6g25t" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.915105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-6g25t" event={"ID":"1b701a42-f649-4657-9250-49188b10d3b9","Type":"ContainerDied","Data":"63f23d24127f696c6ac8a975f8b88fae3bcf7cd62b9d4f19bb8d47d4b8611d9a"} Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.915148 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63f23d24127f696c6ac8a975f8b88fae3bcf7cd62b9d4f19bb8d47d4b8611d9a" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.950625 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.954846 4907 scope.go:117] "RemoveContainer" containerID="4f693ef5bd7b7f8e96bb97b86b9be101cd4aab9df400b462e0a79f5818a5b73b" Mar 13 14:29:11 crc kubenswrapper[4907]: I0313 14:29:11.974002 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58b85ccffc-jt5lv"] Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.069423 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.087517 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.087978 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" containerID="cri-o://7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f" gracePeriod=30 Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.088535 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" containerID="cri-o://7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec" gracePeriod=30 Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.407742 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491309 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491504 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.491666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") pod \"65556420-dc7f-49c8-af59-abafebbb4674\" (UID: \"65556420-dc7f-49c8-af59-abafebbb4674\") " Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.500758 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts" (OuterVolumeSpecName: "scripts") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.507196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2" (OuterVolumeSpecName: "kube-api-access-zvml2") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "kube-api-access-zvml2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.527798 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data" (OuterVolumeSpecName: "config-data") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.538028 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65556420-dc7f-49c8-af59-abafebbb4674" (UID: "65556420-dc7f-49c8-af59-abafebbb4674"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593297 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593321 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593332 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvml2\" (UniqueName: \"kubernetes.io/projected/65556420-dc7f-49c8-af59-abafebbb4674-kube-api-access-zvml2\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.593341 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65556420-dc7f-49c8-af59-abafebbb4674-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.927605 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.927605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-xw7n9" event={"ID":"65556420-dc7f-49c8-af59-abafebbb4674","Type":"ContainerDied","Data":"dc2a57efb2cfb970e90ab542c28a0ad4000fce19809d4298c8e7c3ed553951bb"} Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.927981 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc2a57efb2cfb970e90ab542c28a0ad4000fce19809d4298c8e7c3ed553951bb" Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.928845 4907 generic.go:334] "Generic (PLEG): container finished" podID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerID="7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f" exitCode=143 Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.928908 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerDied","Data":"7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f"} Mar 13 14:29:12 crc kubenswrapper[4907]: I0313 14:29:12.930718 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" containerID="cri-o://0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" gracePeriod=30 Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.014726 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015438 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65556420-dc7f-49c8-af59-abafebbb4674" containerName="nova-cell1-conductor-db-sync" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015465 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="65556420-dc7f-49c8-af59-abafebbb4674" containerName="nova-cell1-conductor-db-sync" Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015484 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b701a42-f649-4657-9250-49188b10d3b9" containerName="nova-manage" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015492 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b701a42-f649-4657-9250-49188b10d3b9" containerName="nova-manage" Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015517 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015525 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" Mar 13 14:29:13 crc kubenswrapper[4907]: E0313 14:29:13.015547 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="init" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015558 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="init" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015844 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b701a42-f649-4657-9250-49188b10d3b9" containerName="nova-manage" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015868 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" containerName="dnsmasq-dns" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.015896 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="65556420-dc7f-49c8-af59-abafebbb4674" containerName="nova-cell1-conductor-db-sync" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.016656 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.018798 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.040345 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.103674 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.103726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.103975 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.205518 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.205976 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.206086 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.210819 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.217094 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.222430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"nova-cell1-conductor-0\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.335250 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.793415 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24" path="/var/lib/kubelet/pods/eed444b5-3ec0-44b6-b0a0-3e3e5a66fe24/volumes" Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.836275 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:29:13 crc kubenswrapper[4907]: I0313 14:29:13.947589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerStarted","Data":"b71e79c4c15864ccbc5c66db3afab587a50c68e66b3d7994a1fc84ed59fd938f"} Mar 13 14:29:14 crc kubenswrapper[4907]: I0313 14:29:14.957341 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerStarted","Data":"77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4"} Mar 13 14:29:14 crc kubenswrapper[4907]: I0313 14:29:14.957741 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:14 crc kubenswrapper[4907]: I0313 14:29:14.977044 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.977027124 podStartE2EDuration="2.977027124s" podCreationTimestamp="2026-03-13 14:29:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:14.96947936 +0000 UTC m=+1453.869267069" watchObservedRunningTime="2026-03-13 14:29:14.977027124 +0000 UTC m=+1453.876814813" Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.210082 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.211894 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.213305 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:15 crc kubenswrapper[4907]: E0313 14:29:15.213347 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:16 crc kubenswrapper[4907]: I0313 14:29:16.976258 4907 generic.go:334] "Generic (PLEG): container finished" podID="636c92f5-234f-4de8-b37b-09be481708e1" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" exitCode=0 Mar 13 14:29:16 crc kubenswrapper[4907]: I0313 14:29:16.976329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerDied","Data":"0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5"} Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.104394 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.189413 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") pod \"636c92f5-234f-4de8-b37b-09be481708e1\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.189607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") pod \"636c92f5-234f-4de8-b37b-09be481708e1\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.189664 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") pod \"636c92f5-234f-4de8-b37b-09be481708e1\" (UID: \"636c92f5-234f-4de8-b37b-09be481708e1\") " Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.195543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc" (OuterVolumeSpecName: "kube-api-access-6c8mc") pod "636c92f5-234f-4de8-b37b-09be481708e1" (UID: "636c92f5-234f-4de8-b37b-09be481708e1"). InnerVolumeSpecName "kube-api-access-6c8mc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.221438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data" (OuterVolumeSpecName: "config-data") pod "636c92f5-234f-4de8-b37b-09be481708e1" (UID: "636c92f5-234f-4de8-b37b-09be481708e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.222903 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "636c92f5-234f-4de8-b37b-09be481708e1" (UID: "636c92f5-234f-4de8-b37b-09be481708e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.292191 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.292237 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/636c92f5-234f-4de8-b37b-09be481708e1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:17 crc kubenswrapper[4907]: I0313 14:29:17.292251 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c8mc\" (UniqueName: \"kubernetes.io/projected/636c92f5-234f-4de8-b37b-09be481708e1-kube-api-access-6c8mc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.014539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.014546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"636c92f5-234f-4de8-b37b-09be481708e1","Type":"ContainerDied","Data":"bc36702cf525d044a0ffd3094a7d069c598da5cdc98d37c88eb61db1ff303142"} Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.014974 4907 scope.go:117] "RemoveContainer" containerID="0307fce5fca271fe0872451ae17b038a302f7cddd1ff8ef3cc5a23fc73b4c7d5" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.021095 4907 generic.go:334] "Generic (PLEG): container finished" podID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerID="7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec" exitCode=0 Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.021128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerDied","Data":"7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec"} Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.102063 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.113934 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.132492 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141406 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.141859 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141896 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.141910 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.141934 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.141941 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142142 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-api" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142172 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="636c92f5-234f-4de8-b37b-09be481708e1" containerName="nova-scheduler-scheduler" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142190 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" containerName="nova-api-log" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.142830 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.152323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.173417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.204718 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.204853 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.204939 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.205009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.206781 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs" (OuterVolumeSpecName: "logs") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.207601 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522d9ac2-33a0-4499-8575-b9e252e1c859-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.229127 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n" (OuterVolumeSpecName: "kube-api-access-2p88n") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "kube-api-access-2p88n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: E0313 14:29:18.232274 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle podName:522d9ac2-33a0-4499-8575-b9e252e1c859 nodeName:}" failed. No retries permitted until 2026-03-13 14:29:18.732225797 +0000 UTC m=+1457.632013486 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859") : error deleting /var/lib/kubelet/pods/522d9ac2-33a0-4499-8575-b9e252e1c859/volume-subpaths: remove /var/lib/kubelet/pods/522d9ac2-33a0-4499-8575-b9e252e1c859/volume-subpaths: no such file or directory Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.234509 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data" (OuterVolumeSpecName: "config-data") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.308816 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.308868 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.308959 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.309065 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.309077 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p88n\" (UniqueName: \"kubernetes.io/projected/522d9ac2-33a0-4499-8575-b9e252e1c859-kube-api-access-2p88n\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.367485 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.367540 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.411288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.411439 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.411471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.414623 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.415367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.432586 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"nova-scheduler-0\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.462957 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.818319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") pod \"522d9ac2-33a0-4499-8575-b9e252e1c859\" (UID: \"522d9ac2-33a0-4499-8575-b9e252e1c859\") " Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.826378 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "522d9ac2-33a0-4499-8575-b9e252e1c859" (UID: "522d9ac2-33a0-4499-8575-b9e252e1c859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.917855 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:18 crc kubenswrapper[4907]: I0313 14:29:18.920550 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522d9ac2-33a0-4499-8575-b9e252e1c859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.031096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerStarted","Data":"a765fc07ea8b46927f8043334da5e0885287349d5eefd55592990b1ae30cfba6"} Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.034824 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522d9ac2-33a0-4499-8575-b9e252e1c859","Type":"ContainerDied","Data":"62294b7ca355e8d04aaa630081ec7abb20dd7d9bb3d9f97d4610e23fae479816"} Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.034857 4907 scope.go:117] "RemoveContainer" containerID="7dd20ac2524e3078e8bda1b7343498987597cfd7e20180905f0e9f71074bc2ec" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.034888 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.080849 4907 scope.go:117] "RemoveContainer" containerID="7f405b45cecbda753272f1e3a9f4ad5fdb7025287b9661064da148b88958c60f" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.106968 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.122519 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.136902 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.138797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.141677 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.160511 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326154 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326384 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326524 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.326981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429054 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429120 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429161 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.429637 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.435274 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.435622 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.450111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"nova-api-0\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.473831 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.816192 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522d9ac2-33a0-4499-8575-b9e252e1c859" path="/var/lib/kubelet/pods/522d9ac2-33a0-4499-8575-b9e252e1c859/volumes" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.817967 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636c92f5-234f-4de8-b37b-09be481708e1" path="/var/lib/kubelet/pods/636c92f5-234f-4de8-b37b-09be481708e1/volumes" Mar 13 14:29:19 crc kubenswrapper[4907]: I0313 14:29:19.936492 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.049458 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerStarted","Data":"7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b"} Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.058301 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerStarted","Data":"11bb2f382efdf360af4e2164320ab74af741c059bdd8ac659bd82b1507cb52af"} Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.082056 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.082029282 podStartE2EDuration="2.082029282s" podCreationTimestamp="2026-03-13 14:29:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:20.078835595 +0000 UTC m=+1458.978623284" watchObservedRunningTime="2026-03-13 14:29:20.082029282 +0000 UTC m=+1458.981816971" Mar 13 14:29:20 crc kubenswrapper[4907]: I0313 14:29:20.970451 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 14:29:21 crc kubenswrapper[4907]: I0313 14:29:21.081850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerStarted","Data":"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6"} Mar 13 14:29:21 crc kubenswrapper[4907]: I0313 14:29:21.082460 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerStarted","Data":"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b"} Mar 13 14:29:21 crc kubenswrapper[4907]: I0313 14:29:21.111187 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.111166431 podStartE2EDuration="2.111166431s" podCreationTimestamp="2026-03-13 14:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:21.105301411 +0000 UTC m=+1460.005089100" watchObservedRunningTime="2026-03-13 14:29:21.111166431 +0000 UTC m=+1460.010954120" Mar 13 14:29:23 crc kubenswrapper[4907]: I0313 14:29:23.380667 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 14:29:23 crc kubenswrapper[4907]: I0313 14:29:23.463592 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.274274 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.274818 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" containerID="cri-o://f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" gracePeriod=30 Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.757683 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.859438 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") pod \"a7f6877c-1e9e-4e17-803d-90efa7d66469\" (UID: \"a7f6877c-1e9e-4e17-803d-90efa7d66469\") " Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.886114 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk" (OuterVolumeSpecName: "kube-api-access-zg7qk") pod "a7f6877c-1e9e-4e17-803d-90efa7d66469" (UID: "a7f6877c-1e9e-4e17-803d-90efa7d66469"). InnerVolumeSpecName "kube-api-access-zg7qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:25 crc kubenswrapper[4907]: I0313 14:29:25.962228 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg7qk\" (UniqueName: \"kubernetes.io/projected/a7f6877c-1e9e-4e17-803d-90efa7d66469-kube-api-access-zg7qk\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138802 4907 generic.go:334] "Generic (PLEG): container finished" podID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" exitCode=2 Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerDied","Data":"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c"} Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a7f6877c-1e9e-4e17-803d-90efa7d66469","Type":"ContainerDied","Data":"3893671130f88340bc9f4b2c1c7ae2e64ebc3c35392832b178f2caf4f860f9fd"} Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138946 4907 scope.go:117] "RemoveContainer" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.138897 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.172126 4907 scope.go:117] "RemoveContainer" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" Mar 13 14:29:26 crc kubenswrapper[4907]: E0313 14:29:26.172745 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c\": container with ID starting with f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c not found: ID does not exist" containerID="f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.172777 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c"} err="failed to get container status \"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c\": rpc error: code = NotFound desc = could not find container \"f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c\": container with ID starting with f5792262fabc2fac4148b46830f3f0bdf740121dbf4200684e158ad3432a779c not found: ID does not exist" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.182420 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.194851 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.201807 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: E0313 14:29:26.202430 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.202459 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.202726 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" containerName="kube-state-metrics" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.203601 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.207117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.210450 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.211945 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.265574 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.265956 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.266012 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.266036 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367028 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367112 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.367226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.373226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.373516 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.376342 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.396338 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"kube-state-metrics-0\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.521653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:29:26 crc kubenswrapper[4907]: I0313 14:29:26.977306 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.150057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerStarted","Data":"50ab44d5af6ebe42498682725ea848a78fc975fa9daf535d39dda1ec402423d1"} Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.249573 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250086 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" containerID="cri-o://62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250152 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" containerID="cri-o://3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250164 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" containerID="cri-o://f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.250127 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" containerID="cri-o://c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" gracePeriod=30 Mar 13 14:29:27 crc kubenswrapper[4907]: I0313 14:29:27.793319 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f6877c-1e9e-4e17-803d-90efa7d66469" path="/var/lib/kubelet/pods/a7f6877c-1e9e-4e17-803d-90efa7d66469/volumes" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165578 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" exitCode=0 Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165921 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" exitCode=2 Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165931 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" exitCode=0 Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.165756 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.166004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.166020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.168374 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerStarted","Data":"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589"} Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.169569 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.190343 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.767998003 podStartE2EDuration="2.190325102s" podCreationTimestamp="2026-03-13 14:29:26 +0000 UTC" firstStartedPulling="2026-03-13 14:29:26.990380324 +0000 UTC m=+1465.890168013" lastFinishedPulling="2026-03-13 14:29:27.412707413 +0000 UTC m=+1466.312495112" observedRunningTime="2026-03-13 14:29:28.185208262 +0000 UTC m=+1467.084995951" watchObservedRunningTime="2026-03-13 14:29:28.190325102 +0000 UTC m=+1467.090112791" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.464108 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 14:29:28 crc kubenswrapper[4907]: I0313 14:29:28.489097 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 14:29:29 crc kubenswrapper[4907]: I0313 14:29:29.222146 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 14:29:29 crc kubenswrapper[4907]: I0313 14:29:29.474900 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:29 crc kubenswrapper[4907]: I0313 14:29:29.475203 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.557082 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.557184 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.819692 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956287 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956380 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.956616 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") pod \"a85017ea-b984-471a-a9a8-22c02d94448b\" (UID: \"a85017ea-b984-471a-a9a8-22c02d94448b\") " Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.958020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.958094 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.968353 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts" (OuterVolumeSpecName: "scripts") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.968495 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb" (OuterVolumeSpecName: "kube-api-access-462xb") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "kube-api-access-462xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:30 crc kubenswrapper[4907]: I0313 14:29:30.991140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.056308 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059229 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-462xb\" (UniqueName: \"kubernetes.io/projected/a85017ea-b984-471a-a9a8-22c02d94448b-kube-api-access-462xb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059256 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059268 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059281 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059290 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.059301 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a85017ea-b984-471a-a9a8-22c02d94448b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.069374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data" (OuterVolumeSpecName: "config-data") pod "a85017ea-b984-471a-a9a8-22c02d94448b" (UID: "a85017ea-b984-471a-a9a8-22c02d94448b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.162640 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85017ea-b984-471a-a9a8-22c02d94448b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210432 4907 generic.go:334] "Generic (PLEG): container finished" podID="a85017ea-b984-471a-a9a8-22c02d94448b" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" exitCode=0 Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01"} Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210508 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a85017ea-b984-471a-a9a8-22c02d94448b","Type":"ContainerDied","Data":"5a88cd9ef6cae1bcbb8c65cf464b690c68ec2dfeb0b4737e783c308ca4a8ffd6"} Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210536 4907 scope.go:117] "RemoveContainer" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.210688 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.253555 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.257530 4907 scope.go:117] "RemoveContainer" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.274347 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.291500 4907 scope.go:117] "RemoveContainer" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293008 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293504 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293527 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293548 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293555 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293572 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293579 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.293601 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293607 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293796 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-notification-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293810 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="sg-core" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293817 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="ceilometer-central-agent" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.293844 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" containerName="proxy-httpd" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.295819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.298800 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.299089 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.299347 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.303051 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.331026 4907 scope.go:117] "RemoveContainer" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.357243 4907 scope.go:117] "RemoveContainer" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.373018 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9\": container with ID starting with c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9 not found: ID does not exist" containerID="c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.373072 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9"} err="failed to get container status \"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9\": rpc error: code = NotFound desc = could not find container \"c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9\": container with ID starting with c2a32ce2fcf985bbc86c96669646554e0cd290d1eaa4396d46e02db24cec19b9 not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.373096 4907 scope.go:117] "RemoveContainer" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.373997 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7\": container with ID starting with f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7 not found: ID does not exist" containerID="f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374020 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7"} err="failed to get container status \"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7\": rpc error: code = NotFound desc = could not find container \"f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7\": container with ID starting with f629b415ac0d40fdc7a871c081a5a0ceecff9578a74f62763cbbaae7d50396c7 not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374035 4907 scope.go:117] "RemoveContainer" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374335 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374431 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374508 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374550 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.374607 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.379316 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01\": container with ID starting with 3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01 not found: ID does not exist" containerID="3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.379361 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01"} err="failed to get container status \"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01\": rpc error: code = NotFound desc = could not find container \"3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01\": container with ID starting with 3b55543c8226645f0d4a440e41c724886657fe5c5d9924185cd3344bad9ddc01 not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.379390 4907 scope.go:117] "RemoveContainer" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" Mar 13 14:29:31 crc kubenswrapper[4907]: E0313 14:29:31.379683 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a\": container with ID starting with 62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a not found: ID does not exist" containerID="62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.379707 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a"} err="failed to get container status \"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a\": rpc error: code = NotFound desc = could not find container \"62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a\": container with ID starting with 62988d506f50133c1068883469efb51974597559ba58796a2aae59bf02bebb3a not found: ID does not exist" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475830 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475908 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475956 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.475994 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476074 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476109 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.476165 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.484849 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.485564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.493343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.494058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.494350 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.494738 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.498163 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.501619 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"ceilometer-0\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.632057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:31 crc kubenswrapper[4907]: I0313 14:29:31.806847 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a85017ea-b984-471a-a9a8-22c02d94448b" path="/var/lib/kubelet/pods/a85017ea-b984-471a-a9a8-22c02d94448b/volumes" Mar 13 14:29:32 crc kubenswrapper[4907]: I0313 14:29:32.102527 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:32 crc kubenswrapper[4907]: I0313 14:29:32.220501 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"a73a14261a590eee8da5263fe0ee191d82682588896d368eb8d1f46e7f9984ba"} Mar 13 14:29:33 crc kubenswrapper[4907]: I0313 14:29:33.234081 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6"} Mar 13 14:29:34 crc kubenswrapper[4907]: I0313 14:29:34.244640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc"} Mar 13 14:29:34 crc kubenswrapper[4907]: I0313 14:29:34.245023 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.220195 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.263397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerStarted","Data":"8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.263812 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.266845 4907 generic.go:334] "Generic (PLEG): container finished" podID="a99d655e-f6db-4c83-a950-b935588b7df1" containerID="ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961" exitCode=137 Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.266932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerDied","Data":"ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.269508 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") pod \"1915a375-806d-43bb-8749-9524db071714\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.269543 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") pod \"1915a375-806d-43bb-8749-9524db071714\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.269587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") pod \"1915a375-806d-43bb-8749-9524db071714\" (UID: \"1915a375-806d-43bb-8749-9524db071714\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271171 4907 generic.go:334] "Generic (PLEG): container finished" podID="1915a375-806d-43bb-8749-9524db071714" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" exitCode=137 Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271378 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerDied","Data":"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1915a375-806d-43bb-8749-9524db071714","Type":"ContainerDied","Data":"878ef11d68a7add33ce200311e92840ebc8d43565aea3bfc5d4ea32961e387a8"} Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271421 4907 scope.go:117] "RemoveContainer" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.271581 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.275973 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq" (OuterVolumeSpecName: "kube-api-access-phktq") pod "1915a375-806d-43bb-8749-9524db071714" (UID: "1915a375-806d-43bb-8749-9524db071714"). InnerVolumeSpecName "kube-api-access-phktq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.298381 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.084994597 podStartE2EDuration="5.298343132s" podCreationTimestamp="2026-03-13 14:29:31 +0000 UTC" firstStartedPulling="2026-03-13 14:29:32.118316015 +0000 UTC m=+1471.018103704" lastFinishedPulling="2026-03-13 14:29:35.33166455 +0000 UTC m=+1474.231452239" observedRunningTime="2026-03-13 14:29:36.281383231 +0000 UTC m=+1475.181170920" watchObservedRunningTime="2026-03-13 14:29:36.298343132 +0000 UTC m=+1475.198130821" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.305373 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1915a375-806d-43bb-8749-9524db071714" (UID: "1915a375-806d-43bb-8749-9524db071714"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.306269 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data" (OuterVolumeSpecName: "config-data") pod "1915a375-806d-43bb-8749-9524db071714" (UID: "1915a375-806d-43bb-8749-9524db071714"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.314933 4907 scope.go:117] "RemoveContainer" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" Mar 13 14:29:36 crc kubenswrapper[4907]: E0313 14:29:36.315726 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285\": container with ID starting with 45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285 not found: ID does not exist" containerID="45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.315798 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285"} err="failed to get container status \"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285\": rpc error: code = NotFound desc = could not find container \"45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285\": container with ID starting with 45b20ff335a1fe9f36ec78c46577e6fb80b04fdde4e9de1eb52326676228f285 not found: ID does not exist" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.373497 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phktq\" (UniqueName: \"kubernetes.io/projected/1915a375-806d-43bb-8749-9524db071714-kube-api-access-phktq\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.373550 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.373569 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1915a375-806d-43bb-8749-9524db071714-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.540983 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.652840 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.672518 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.682001 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: E0313 14:29:36.682319 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.682330 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.682527 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1915a375-806d-43bb-8749-9524db071714" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.683148 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.690703 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.691920 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.693866 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.713347 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.728427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.885279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.885749 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.886773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.886932 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") pod \"a99d655e-f6db-4c83-a950-b935588b7df1\" (UID: \"a99d655e-f6db-4c83-a950-b935588b7df1\") " Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.887124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs" (OuterVolumeSpecName: "logs") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888036 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888139 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888218 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888353 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.888656 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a99d655e-f6db-4c83-a950-b935588b7df1-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.891760 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s" (OuterVolumeSpecName: "kube-api-access-v4n9s") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "kube-api-access-v4n9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.922229 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.954135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data" (OuterVolumeSpecName: "config-data") pod "a99d655e-f6db-4c83-a950-b935588b7df1" (UID: "a99d655e-f6db-4c83-a950-b935588b7df1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990161 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990251 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990316 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990396 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990502 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990519 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4n9s\" (UniqueName: \"kubernetes.io/projected/a99d655e-f6db-4c83-a950-b935588b7df1-kube-api-access-v4n9s\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.990532 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a99d655e-f6db-4c83-a950-b935588b7df1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.993564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.993599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:36 crc kubenswrapper[4907]: I0313 14:29:36.994984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.001428 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.007029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"nova-cell1-novncproxy-0\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.039486 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.284309 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a99d655e-f6db-4c83-a950-b935588b7df1","Type":"ContainerDied","Data":"f5827c9400a454954bce2169d1bdd16e49d4ec2488819960057446785871a720"} Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.284582 4907 scope.go:117] "RemoveContainer" containerID="ab262217f9b17cb1ba1a68c95414fc65431f7517928e9e7fc8f78902f26ff961" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.284711 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.328209 4907 scope.go:117] "RemoveContainer" containerID="c48900e9ab50388f765d2b50bcf7cb048feb1b78e5669c08c203f15509de3dac" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.332058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.347005 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.358476 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: E0313 14:29:37.358986 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359010 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" Mar 13 14:29:37 crc kubenswrapper[4907]: E0313 14:29:37.359055 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359063 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359256 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-log" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.359287 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" containerName="nova-metadata-metadata" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.360442 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.363327 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.363526 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.363769 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.397792 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.397954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.397981 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.398067 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.398111 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.474908 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.475909 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.481060 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501161 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501258 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501407 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501433 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.501611 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.514862 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.514980 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.515000 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.524051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"nova-metadata-0\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.718732 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.795275 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1915a375-806d-43bb-8749-9524db071714" path="/var/lib/kubelet/pods/1915a375-806d-43bb-8749-9524db071714/volumes" Mar 13 14:29:37 crc kubenswrapper[4907]: I0313 14:29:37.795971 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a99d655e-f6db-4c83-a950-b935588b7df1" path="/var/lib/kubelet/pods/a99d655e-f6db-4c83-a950-b935588b7df1/volumes" Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.184626 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:38 crc kubenswrapper[4907]: W0313 14:29:38.186281 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39488342_aa1a_4dba_bd5f_556da5081904.slice/crio-a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a WatchSource:0}: Error finding container a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a: Status 404 returned error can't find the container with id a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.306240 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerStarted","Data":"9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7"} Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.306279 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerStarted","Data":"1cd31d9d43a5e6a87bb0836d84499cc5b9fab420b814ef48408f1a9549204fd0"} Mar 13 14:29:38 crc kubenswrapper[4907]: I0313 14:29:38.310485 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerStarted","Data":"a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a"} Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.324103 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerStarted","Data":"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143"} Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.324464 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerStarted","Data":"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507"} Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.344710 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.344687712 podStartE2EDuration="3.344687712s" podCreationTimestamp="2026-03-13 14:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:38.330523681 +0000 UTC m=+1477.230311380" watchObservedRunningTime="2026-03-13 14:29:39.344687712 +0000 UTC m=+1478.244475401" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.478070 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.483026 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.484367 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:29:39 crc kubenswrapper[4907]: I0313 14:29:39.508975 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.508954433 podStartE2EDuration="2.508954433s" podCreationTimestamp="2026-03-13 14:29:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:39.342893894 +0000 UTC m=+1478.242681593" watchObservedRunningTime="2026-03-13 14:29:39.508954433 +0000 UTC m=+1478.408742132" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.334228 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.512691 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.514306 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.541338 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670188 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670262 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670402 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670597 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.670722 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772569 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772760 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772805 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772841 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.772907 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.773873 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.774029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.774484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.775139 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.775155 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.794743 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"dnsmasq-dns-fdb8f6449-nb4km\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:40 crc kubenswrapper[4907]: I0313 14:29:40.839870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:41 crc kubenswrapper[4907]: W0313 14:29:41.998040 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7da13722_732a_4b65_a894_fe4612f30d75.slice/crio-fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929 WatchSource:0}: Error finding container fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929: Status 404 returned error can't find the container with id fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929 Mar 13 14:29:41 crc kubenswrapper[4907]: I0313 14:29:41.999305 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.040206 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.354629 4907 generic.go:334] "Generic (PLEG): container finished" podID="7da13722-732a-4b65-a894-fe4612f30d75" containerID="9c33ada0f7c03cde714537a63faadeb180f6ac07dd15e3f4ce8c7087574733bb" exitCode=0 Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.354696 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerDied","Data":"9c33ada0f7c03cde714537a63faadeb180f6ac07dd15e3f4ce8c7087574733bb"} Mar 13 14:29:42 crc kubenswrapper[4907]: I0313 14:29:42.354774 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerStarted","Data":"fbaac4295e0e2ca7122793d1f083ea125d689fae676b037de247b0fa68fc7929"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.196788 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197337 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" containerID="cri-o://9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197401 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" containerID="cri-o://8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197457 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" containerID="cri-o://5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.197496 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" containerID="cri-o://5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.367698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerStarted","Data":"890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.368544 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.375954 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538" exitCode=0 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.375987 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc" exitCode=2 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.376010 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.376033 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc"} Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.393580 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" podStartSLOduration=3.393558399 podStartE2EDuration="3.393558399s" podCreationTimestamp="2026-03-13 14:29:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:43.386312332 +0000 UTC m=+1482.286100021" watchObservedRunningTime="2026-03-13 14:29:43.393558399 +0000 UTC m=+1482.293346088" Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.443715 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.444038 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" containerID="cri-o://97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" gracePeriod=30 Mar 13 14:29:43 crc kubenswrapper[4907]: I0313 14:29:43.444122 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" containerID="cri-o://bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" gracePeriod=30 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.391567 4907 generic.go:334] "Generic (PLEG): container finished" podID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" exitCode=143 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.391936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerDied","Data":"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b"} Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.394717 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515" exitCode=0 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.394741 4907 generic.go:334] "Generic (PLEG): container finished" podID="41167fb8-73ed-4487-975c-365854b0d586" containerID="9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6" exitCode=0 Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.395676 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515"} Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.395707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6"} Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.632149 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759017 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759129 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759150 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759250 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759333 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759369 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759443 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759845 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.759956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") pod \"41167fb8-73ed-4487-975c-365854b0d586\" (UID: \"41167fb8-73ed-4487-975c-365854b0d586\") " Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.760740 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.760828 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41167fb8-73ed-4487-975c-365854b0d586-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.766119 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts" (OuterVolumeSpecName: "scripts") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.766439 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5" (OuterVolumeSpecName: "kube-api-access-w4ww5") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "kube-api-access-w4ww5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.802360 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.811400 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.849076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.858942 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data" (OuterVolumeSpecName: "config-data") pod "41167fb8-73ed-4487-975c-365854b0d586" (UID: "41167fb8-73ed-4487-975c-365854b0d586"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862431 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862477 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862492 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862506 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4ww5\" (UniqueName: \"kubernetes.io/projected/41167fb8-73ed-4487-975c-365854b0d586-kube-api-access-w4ww5\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862518 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:44 crc kubenswrapper[4907]: I0313 14:29:44.862533 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41167fb8-73ed-4487-975c-365854b0d586-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.405830 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41167fb8-73ed-4487-975c-365854b0d586","Type":"ContainerDied","Data":"a73a14261a590eee8da5263fe0ee191d82682588896d368eb8d1f46e7f9984ba"} Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.405894 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.406189 4907 scope.go:117] "RemoveContainer" containerID="8e591aebdcd4fb21580fb8235da93dec31fd3679a3943e9171c4f1541d57c538" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.430432 4907 scope.go:117] "RemoveContainer" containerID="5939f8118a1b021358a2a1f3def6573f42d731f7807d9451cfe9c468306f6dbc" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.447120 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.456600 4907 scope.go:117] "RemoveContainer" containerID="5ecb51a13db91baac3a9eb0c7577b30ef982b63567b8d459d38ac04dfa9a4515" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.464032 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.473794 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474153 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474169 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474190 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474198 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474207 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474213 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" Mar 13 14:29:45 crc kubenswrapper[4907]: E0313 14:29:45.474237 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474243 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474404 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="proxy-httpd" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474423 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-central-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474445 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="sg-core" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.474458 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41167fb8-73ed-4487-975c-365854b0d586" containerName="ceilometer-notification-agent" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.476201 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.479564 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.479744 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.479774 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.482091 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.482115 4907 scope.go:117] "RemoveContainer" containerID="9b5ab98c8fa734605e5e5d3d4bc91ad54650c1db1f1d0d56f2cbf2225fe14db6" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584112 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584207 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584285 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584349 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584380 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.584460 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686604 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686692 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686710 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686894 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.686943 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.687779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.688165 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.692371 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.694150 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.695103 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.696210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.697536 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.712202 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"ceilometer-0\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " pod="openstack/ceilometer-0" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.793080 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41167fb8-73ed-4487-975c-365854b0d586" path="/var/lib/kubelet/pods/41167fb8-73ed-4487-975c-365854b0d586/volumes" Mar 13 14:29:45 crc kubenswrapper[4907]: I0313 14:29:45.796723 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:29:46 crc kubenswrapper[4907]: I0313 14:29:46.587016 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.040542 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.062962 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.189592 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320190 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320601 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320820 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.320859 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") pod \"41063592-cda5-443f-87e3-20c2c6ec5c73\" (UID: \"41063592-cda5-443f-87e3-20c2c6ec5c73\") " Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.321224 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs" (OuterVolumeSpecName: "logs") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.321530 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41063592-cda5-443f-87e3-20c2c6ec5c73-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.326436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp" (OuterVolumeSpecName: "kube-api-access-zhhbp") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "kube-api-access-zhhbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.355755 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data" (OuterVolumeSpecName: "config-data") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.376053 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41063592-cda5-443f-87e3-20c2c6ec5c73" (UID: "41063592-cda5-443f-87e3-20c2c6ec5c73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.423232 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.423261 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhhbp\" (UniqueName: \"kubernetes.io/projected/41063592-cda5-443f-87e3-20c2c6ec5c73-kube-api-access-zhhbp\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.423270 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41063592-cda5-443f-87e3-20c2c6ec5c73-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.555129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.555181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"b4ef9b58ba405d6fb537a91498e7353e4bf4deae5a0e5f61d8548bb330be20cb"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557606 4907 generic.go:334] "Generic (PLEG): container finished" podID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" exitCode=0 Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557655 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557678 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerDied","Data":"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557748 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"41063592-cda5-443f-87e3-20c2c6ec5c73","Type":"ContainerDied","Data":"11bb2f382efdf360af4e2164320ab74af741c059bdd8ac659bd82b1507cb52af"} Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.557771 4907 scope.go:117] "RemoveContainer" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.586297 4907 scope.go:117] "RemoveContainer" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.590171 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.672264 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.676575 4907 scope.go:117] "RemoveContainer" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.678721 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6\": container with ID starting with bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6 not found: ID does not exist" containerID="bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.678775 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6"} err="failed to get container status \"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6\": rpc error: code = NotFound desc = could not find container \"bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6\": container with ID starting with bee51476e0503a2fc3a96ca20f1d42b6cb92e1822245b7b21033273b10f0e7a6 not found: ID does not exist" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.678807 4907 scope.go:117] "RemoveContainer" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.680181 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b\": container with ID starting with 97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b not found: ID does not exist" containerID="97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.680226 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b"} err="failed to get container status \"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b\": rpc error: code = NotFound desc = could not find container \"97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b\": container with ID starting with 97742dae0519818b9c495c23cd7ee1fe57412023b1fc525ca01e7fc2236cef2b not found: ID does not exist" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.692955 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.707958 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.708585 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.708622 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" Mar 13 14:29:47 crc kubenswrapper[4907]: E0313 14:29:47.708651 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.708658 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.708983 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-api" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.709032 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" containerName="nova-api-log" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.710424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.713654 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.714151 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.714547 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.719449 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.719495 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.721645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.811948 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41063592-cda5-443f-87e3-20c2c6ec5c73" path="/var/lib/kubelet/pods/41063592-cda5-443f-87e3-20c2c6ec5c73/volumes" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832375 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832419 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832581 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.832600 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.874472 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.893258 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.897854 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.898122 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.933960 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934672 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.934970 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.935067 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.938711 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.942589 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.943348 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.944088 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.952386 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:47 crc kubenswrapper[4907]: I0313 14:29:47.959556 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:47.989080 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"nova-api-0\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " pod="openstack/nova-api-0" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037022 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037074 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037205 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.037243 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.065847 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139630 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.139648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.149429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.150435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.152441 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.159037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"nova-cell1-cell-mapping-2pvp9\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.358183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.578171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04"} Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.635562 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.734113 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.734142 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.206:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:29:48 crc kubenswrapper[4907]: I0313 14:29:48.839785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.589174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerStarted","Data":"714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.589540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerStarted","Data":"9a7bf9979b09ab5543692c5fb89940c7553ed9a8ca26785e17126dc549ce6605"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.591999 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerStarted","Data":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.592115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerStarted","Data":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.592175 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerStarted","Data":"2712a8612ce027f409c3dbd5c6adbabefb48875be5bd50702de7d61655bda2a8"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.598397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5"} Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.627675 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2pvp9" podStartSLOduration=2.627651369 podStartE2EDuration="2.627651369s" podCreationTimestamp="2026-03-13 14:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:49.6114689 +0000 UTC m=+1488.511256589" watchObservedRunningTime="2026-03-13 14:29:49.627651369 +0000 UTC m=+1488.527439068" Mar 13 14:29:49 crc kubenswrapper[4907]: I0313 14:29:49.640168 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.640149538 podStartE2EDuration="2.640149538s" podCreationTimestamp="2026-03-13 14:29:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:49.638139594 +0000 UTC m=+1488.537927293" watchObservedRunningTime="2026-03-13 14:29:49.640149538 +0000 UTC m=+1488.539937227" Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.610648 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerStarted","Data":"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c"} Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.631721 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9883929930000002 podStartE2EDuration="5.631697206s" podCreationTimestamp="2026-03-13 14:29:45 +0000 UTC" firstStartedPulling="2026-03-13 14:29:46.582551852 +0000 UTC m=+1485.482339541" lastFinishedPulling="2026-03-13 14:29:50.225856065 +0000 UTC m=+1489.125643754" observedRunningTime="2026-03-13 14:29:50.62960675 +0000 UTC m=+1489.529394439" watchObservedRunningTime="2026-03-13 14:29:50.631697206 +0000 UTC m=+1489.531484895" Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.841326 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.905993 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:50 crc kubenswrapper[4907]: I0313 14:29:50.907070 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" containerID="cri-o://215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" gracePeriod=10 Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.441281 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.510828 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.510988 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511111 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511154 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511181 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.511221 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") pod \"f56f7387-a1c9-4250-8abf-6c0a3830970e\" (UID: \"f56f7387-a1c9-4250-8abf-6c0a3830970e\") " Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.518112 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc" (OuterVolumeSpecName: "kube-api-access-8c8lc") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "kube-api-access-8c8lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.582800 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.589015 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.595313 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.605332 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613040 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613077 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613095 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c8lc\" (UniqueName: \"kubernetes.io/projected/f56f7387-a1c9-4250-8abf-6c0a3830970e-kube-api-access-8c8lc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613109 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.613121 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.614235 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config" (OuterVolumeSpecName: "config") pod "f56f7387-a1c9-4250-8abf-6c0a3830970e" (UID: "f56f7387-a1c9-4250-8abf-6c0a3830970e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.621653 4907 generic.go:334] "Generic (PLEG): container finished" podID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" exitCode=0 Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.623000 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.627949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerDied","Data":"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a"} Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.628007 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.628019 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69b4446475-ttj9k" event={"ID":"f56f7387-a1c9-4250-8abf-6c0a3830970e","Type":"ContainerDied","Data":"c42518fd3cb49ffcc3d612c9faabeba7f93ee461bc56bb4d3b0a86fdeb96cbc5"} Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.628041 4907 scope.go:117] "RemoveContainer" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.655742 4907 scope.go:117] "RemoveContainer" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.665920 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.674891 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69b4446475-ttj9k"] Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.697750 4907 scope.go:117] "RemoveContainer" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" Mar 13 14:29:51 crc kubenswrapper[4907]: E0313 14:29:51.698162 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a\": container with ID starting with 215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a not found: ID does not exist" containerID="215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.698190 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a"} err="failed to get container status \"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a\": rpc error: code = NotFound desc = could not find container \"215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a\": container with ID starting with 215938dfe070eab04984e52eb7b82e6d3f56a88ab88790cd28ad3b1a4bc7d77a not found: ID does not exist" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.698213 4907 scope.go:117] "RemoveContainer" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" Mar 13 14:29:51 crc kubenswrapper[4907]: E0313 14:29:51.698401 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624\": container with ID starting with d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624 not found: ID does not exist" containerID="d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.698418 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624"} err="failed to get container status \"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624\": rpc error: code = NotFound desc = could not find container \"d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624\": container with ID starting with d71398ef9c9b9dc145195170a4e15d1e48bb6494190d13130bb61281addfc624 not found: ID does not exist" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.714983 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f56f7387-a1c9-4250-8abf-6c0a3830970e-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:51 crc kubenswrapper[4907]: I0313 14:29:51.805055 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" path="/var/lib/kubelet/pods/f56f7387-a1c9-4250-8abf-6c0a3830970e/volumes" Mar 13 14:29:54 crc kubenswrapper[4907]: I0313 14:29:54.650253 4907 generic.go:334] "Generic (PLEG): container finished" podID="8901c771-b45a-4147-9f96-cef784165281" containerID="714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee" exitCode=0 Mar 13 14:29:54 crc kubenswrapper[4907]: I0313 14:29:54.650344 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerDied","Data":"714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee"} Mar 13 14:29:55 crc kubenswrapper[4907]: I0313 14:29:55.719293 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:55 crc kubenswrapper[4907]: I0313 14:29:55.720625 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.026410 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.099867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.100135 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.100357 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.100571 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") pod \"8901c771-b45a-4147-9f96-cef784165281\" (UID: \"8901c771-b45a-4147-9f96-cef784165281\") " Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.105941 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts" (OuterVolumeSpecName: "scripts") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.106202 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx" (OuterVolumeSpecName: "kube-api-access-h9nqx") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "kube-api-access-h9nqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.127369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.132698 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data" (OuterVolumeSpecName: "config-data") pod "8901c771-b45a-4147-9f96-cef784165281" (UID: "8901c771-b45a-4147-9f96-cef784165281"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.204970 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9nqx\" (UniqueName: \"kubernetes.io/projected/8901c771-b45a-4147-9f96-cef784165281-kube-api-access-h9nqx\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.205204 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.205260 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.205330 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8901c771-b45a-4147-9f96-cef784165281-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.669464 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2pvp9" event={"ID":"8901c771-b45a-4147-9f96-cef784165281","Type":"ContainerDied","Data":"9a7bf9979b09ab5543692c5fb89940c7553ed9a8ca26785e17126dc549ce6605"} Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.669819 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a7bf9979b09ab5543692c5fb89940c7553ed9a8ca26785e17126dc549ce6605" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.669518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2pvp9" Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.848549 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.848897 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" containerID="cri-o://c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" gracePeriod=30 Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.849054 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" containerID="cri-o://8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" gracePeriod=30 Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.938900 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.939133 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" containerID="cri-o://7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" gracePeriod=30 Mar 13 14:29:56 crc kubenswrapper[4907]: I0313 14:29:56.954467 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.385403 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528429 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528510 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528547 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528633 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.528722 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") pod \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\" (UID: \"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e\") " Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.533934 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk" (OuterVolumeSpecName: "kube-api-access-q8tvk") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "kube-api-access-q8tvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.534305 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs" (OuterVolumeSpecName: "logs") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.561663 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.561977 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data" (OuterVolumeSpecName: "config-data") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.596527 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.598630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" (UID: "4851fd7f-a2e6-4ec1-96ea-7c0086edd03e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631188 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631228 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631238 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631251 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8tvk\" (UniqueName: \"kubernetes.io/projected/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-kube-api-access-q8tvk\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631263 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.631272 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678832 4907 generic.go:334] "Generic (PLEG): container finished" podID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" exitCode=0 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678867 4907 generic.go:334] "Generic (PLEG): container finished" podID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" exitCode=143 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678904 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678903 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerDied","Data":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678959 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerDied","Data":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4851fd7f-a2e6-4ec1-96ea-7c0086edd03e","Type":"ContainerDied","Data":"2712a8612ce027f409c3dbd5c6adbabefb48875be5bd50702de7d61655bda2a8"} Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.678993 4907 scope.go:117] "RemoveContainer" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.679512 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" containerID="cri-o://ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" gracePeriod=30 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.679633 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" containerID="cri-o://3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" gracePeriod=30 Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.705760 4907 scope.go:117] "RemoveContainer" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.720301 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.729398 4907 scope.go:117] "RemoveContainer" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.730042 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": container with ID starting with 8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b not found: ID does not exist" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730096 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} err="failed to get container status \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": rpc error: code = NotFound desc = could not find container \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": container with ID starting with 8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730123 4907 scope.go:117] "RemoveContainer" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.730544 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": container with ID starting with c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53 not found: ID does not exist" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730581 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} err="failed to get container status \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": rpc error: code = NotFound desc = could not find container \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": container with ID starting with c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53 not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730608 4907 scope.go:117] "RemoveContainer" containerID="8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730925 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b"} err="failed to get container status \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": rpc error: code = NotFound desc = could not find container \"8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b\": container with ID starting with 8a9e8c8e4a79ece75fdb50c32528e4d6f87181c842fcd7d1b9802e179e72bd3b not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.730944 4907 scope.go:117] "RemoveContainer" containerID="c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.731279 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53"} err="failed to get container status \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": rpc error: code = NotFound desc = could not find container \"c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53\": container with ID starting with c4919269912183df77206bdb4ba3361c6048b9cadeaad50a6bcce3ec26e11c53 not found: ID does not exist" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.732554 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.743698 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744168 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8901c771-b45a-4147-9f96-cef784165281" containerName="nova-manage" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744189 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8901c771-b45a-4147-9f96-cef784165281" containerName="nova-manage" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744203 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744220 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744233 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744242 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744258 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744266 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" Mar 13 14:29:57 crc kubenswrapper[4907]: E0313 14:29:57.744292 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="init" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744301 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="init" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744523 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-log" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744553 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56f7387-a1c9-4250-8abf-6c0a3830970e" containerName="dnsmasq-dns" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744568 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" containerName="nova-api-api" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.744585 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8901c771-b45a-4147-9f96-cef784165281" containerName="nova-manage" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.745653 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.749612 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.750640 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.750765 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.757937 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.811005 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4851fd7f-a2e6-4ec1-96ea-7c0086edd03e" path="/var/lib/kubelet/pods/4851fd7f-a2e6-4ec1-96ea-7c0086edd03e/volumes" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834639 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834789 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.834923 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936498 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936548 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936652 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.936925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.941507 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.942100 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.942116 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.942749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:57 crc kubenswrapper[4907]: I0313 14:29:57.958326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"nova-api-0\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " pod="openstack/nova-api-0" Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.062931 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.466464 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.468166 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.469676 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:29:58 crc kubenswrapper[4907]: E0313 14:29:58.469707 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.514134 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.691989 4907 generic.go:334] "Generic (PLEG): container finished" podID="39488342-aa1a-4dba-bd5f-556da5081904" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" exitCode=143 Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.692069 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerDied","Data":"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507"} Mar 13 14:29:58 crc kubenswrapper[4907]: I0313 14:29:58.693786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerStarted","Data":"ed58fd3bb4c7be310b9f67af6133d99b09325a11963ab01983393f4f0a20f895"} Mar 13 14:29:59 crc kubenswrapper[4907]: I0313 14:29:59.710368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerStarted","Data":"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd"} Mar 13 14:29:59 crc kubenswrapper[4907]: I0313 14:29:59.710416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerStarted","Data":"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd"} Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.134430 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.134406004 podStartE2EDuration="3.134406004s" podCreationTimestamp="2026-03-13 14:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:29:59.757490563 +0000 UTC m=+1498.657278252" watchObservedRunningTime="2026-03-13 14:30:00.134406004 +0000 UTC m=+1499.034193703" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.145122 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.146565 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153097 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153227 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153310 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.153795 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.155292 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.160274 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.160538 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.164569 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.184473 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284542 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.284986 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"auto-csr-approver-29556870-8mlkz\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386798 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"auto-csr-approver-29556870-8mlkz\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.386944 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.388070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.395535 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.403982 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"collect-profiles-29556870-rdhxg\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.404670 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"auto-csr-approver-29556870-8mlkz\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.475902 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.495219 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.964274 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:30:00 crc kubenswrapper[4907]: W0313 14:30:00.966816 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7993f31_1072_4c83_a3ac_16dca5eb04a3.slice/crio-0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da WatchSource:0}: Error finding container 0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da: Status 404 returned error can't find the container with id 0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da Mar 13 14:30:00 crc kubenswrapper[4907]: I0313 14:30:00.972390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 14:30:00 crc kubenswrapper[4907]: W0313 14:30:00.983134 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf9d29dd_6749_4fcd_82a8_ea528d832fcc.slice/crio-7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224 WatchSource:0}: Error finding container 7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224: Status 404 returned error can't find the container with id 7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224 Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.227672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303237 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303417 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.303676 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") pod \"39488342-aa1a-4dba-bd5f-556da5081904\" (UID: \"39488342-aa1a-4dba-bd5f-556da5081904\") " Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.304109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs" (OuterVolumeSpecName: "logs") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.304378 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39488342-aa1a-4dba-bd5f-556da5081904-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.322916 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9" (OuterVolumeSpecName: "kube-api-access-774w9") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "kube-api-access-774w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.343956 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.347134 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data" (OuterVolumeSpecName: "config-data") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.373162 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "39488342-aa1a-4dba-bd5f-556da5081904" (UID: "39488342-aa1a-4dba-bd5f-556da5081904"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405663 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405702 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-774w9\" (UniqueName: \"kubernetes.io/projected/39488342-aa1a-4dba-bd5f-556da5081904-kube-api-access-774w9\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405720 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.405730 4907 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39488342-aa1a-4dba-bd5f-556da5081904-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.732098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" event={"ID":"e7993f31-1072-4c83-a3ac-16dca5eb04a3","Type":"ContainerStarted","Data":"0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734782 4907 generic.go:334] "Generic (PLEG): container finished" podID="39488342-aa1a-4dba-bd5f-556da5081904" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" exitCode=0 Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerDied","Data":"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734904 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39488342-aa1a-4dba-bd5f-556da5081904","Type":"ContainerDied","Data":"a7aca03c6cc93dfbda04afd1635961a9edebe57ed61e75bfcd08d8cd3550955a"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734916 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.734934 4907 scope.go:117] "RemoveContainer" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.739560 4907 generic.go:334] "Generic (PLEG): container finished" podID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerID="46dfd46e5c29fa7764c024392038bd85a0e04092c9a231bbd08b67969852ba5c" exitCode=0 Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.739609 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" event={"ID":"af9d29dd-6749-4fcd-82a8-ea528d832fcc","Type":"ContainerDied","Data":"46dfd46e5c29fa7764c024392038bd85a0e04092c9a231bbd08b67969852ba5c"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.739636 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" event={"ID":"af9d29dd-6749-4fcd-82a8-ea528d832fcc","Type":"ContainerStarted","Data":"7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224"} Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.767099 4907 scope.go:117] "RemoveContainer" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.795772 4907 scope.go:117] "RemoveContainer" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.796235 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143\": container with ID starting with 3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143 not found: ID does not exist" containerID="3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.796265 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143"} err="failed to get container status \"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143\": rpc error: code = NotFound desc = could not find container \"3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143\": container with ID starting with 3f0638d25e0be962bf528749f86e7ed7d982188df63f11eef31f67ffdeb4e143 not found: ID does not exist" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.796283 4907 scope.go:117] "RemoveContainer" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.796533 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507\": container with ID starting with ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507 not found: ID does not exist" containerID="ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.796567 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507"} err="failed to get container status \"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507\": rpc error: code = NotFound desc = could not find container \"ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507\": container with ID starting with ee178515d58e37a284bcf6959726bc0672f424f0bddc3d61b496fcc0bf679507 not found: ID does not exist" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.823827 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.823911 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.823931 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.824559 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.824581 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" Mar 13 14:30:01 crc kubenswrapper[4907]: E0313 14:30:01.824639 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.824649 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.825017 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-metadata" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.825067 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="39488342-aa1a-4dba-bd5f-556da5081904" containerName="nova-metadata-log" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.830259 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.832586 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.833117 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.846542 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923034 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923329 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923493 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923576 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:01 crc kubenswrapper[4907]: I0313 14:30:01.923729 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026662 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026824 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026848 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.026863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.030841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.044353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.044414 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.049113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.049212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.158894 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.622217 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:02 crc kubenswrapper[4907]: W0313 14:30:02.630541 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e41272f_6e7b_47a7_825e_3571f6a1fd07.slice/crio-68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3 WatchSource:0}: Error finding container 68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3: Status 404 returned error can't find the container with id 68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3 Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.751213 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerID="93b983c69ed0cdc38b2c77a7998230bc3acd8a611b28bc2ce28fc6ae6eacb20d" exitCode=0 Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.751653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" event={"ID":"e7993f31-1072-4c83-a3ac-16dca5eb04a3","Type":"ContainerDied","Data":"93b983c69ed0cdc38b2c77a7998230bc3acd8a611b28bc2ce28fc6ae6eacb20d"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.755242 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerStarted","Data":"68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.756779 4907 generic.go:334] "Generic (PLEG): container finished" podID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" exitCode=0 Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.756977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerDied","Data":"7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.757009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c0304716-7acd-4a2c-a018-73fc4bcbad51","Type":"ContainerDied","Data":"a765fc07ea8b46927f8043334da5e0885287349d5eefd55592990b1ae30cfba6"} Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.757024 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a765fc07ea8b46927f8043334da5e0885287349d5eefd55592990b1ae30cfba6" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.838928 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.948018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") pod \"c0304716-7acd-4a2c-a018-73fc4bcbad51\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.948096 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") pod \"c0304716-7acd-4a2c-a018-73fc4bcbad51\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.948226 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") pod \"c0304716-7acd-4a2c-a018-73fc4bcbad51\" (UID: \"c0304716-7acd-4a2c-a018-73fc4bcbad51\") " Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.952141 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6" (OuterVolumeSpecName: "kube-api-access-xhvp6") pod "c0304716-7acd-4a2c-a018-73fc4bcbad51" (UID: "c0304716-7acd-4a2c-a018-73fc4bcbad51"). InnerVolumeSpecName "kube-api-access-xhvp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:02 crc kubenswrapper[4907]: I0313 14:30:02.976701 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0304716-7acd-4a2c-a018-73fc4bcbad51" (UID: "c0304716-7acd-4a2c-a018-73fc4bcbad51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.017142 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data" (OuterVolumeSpecName: "config-data") pod "c0304716-7acd-4a2c-a018-73fc4bcbad51" (UID: "c0304716-7acd-4a2c-a018-73fc4bcbad51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.051219 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.051521 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhvp6\" (UniqueName: \"kubernetes.io/projected/c0304716-7acd-4a2c-a018-73fc4bcbad51-kube-api-access-xhvp6\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.051537 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0304716-7acd-4a2c-a018-73fc4bcbad51-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.078594 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.152638 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") pod \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.152867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") pod \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.152992 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") pod \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\" (UID: \"af9d29dd-6749-4fcd-82a8-ea528d832fcc\") " Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.154603 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume" (OuterVolumeSpecName: "config-volume") pod "af9d29dd-6749-4fcd-82a8-ea528d832fcc" (UID: "af9d29dd-6749-4fcd-82a8-ea528d832fcc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.159132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "af9d29dd-6749-4fcd-82a8-ea528d832fcc" (UID: "af9d29dd-6749-4fcd-82a8-ea528d832fcc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.165539 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z" (OuterVolumeSpecName: "kube-api-access-d6z4z") pod "af9d29dd-6749-4fcd-82a8-ea528d832fcc" (UID: "af9d29dd-6749-4fcd-82a8-ea528d832fcc"). InnerVolumeSpecName "kube-api-access-d6z4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.260621 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af9d29dd-6749-4fcd-82a8-ea528d832fcc-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.260683 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af9d29dd-6749-4fcd-82a8-ea528d832fcc-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.260695 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6z4z\" (UniqueName: \"kubernetes.io/projected/af9d29dd-6749-4fcd-82a8-ea528d832fcc-kube-api-access-d6z4z\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.766771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" event={"ID":"af9d29dd-6749-4fcd-82a8-ea528d832fcc","Type":"ContainerDied","Data":"7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224"} Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.766815 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c2a8f78d2f024990112bf064e09dd2f29128e2c064da8d891a6688df0ffe224" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.766828 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.769098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerStarted","Data":"ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e"} Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.769145 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.769157 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerStarted","Data":"c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d"} Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.826793 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.826771254 podStartE2EDuration="2.826771254s" podCreationTimestamp="2026-03-13 14:30:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:30:03.797093632 +0000 UTC m=+1502.696881331" watchObservedRunningTime="2026-03-13 14:30:03.826771254 +0000 UTC m=+1502.726558943" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.842692 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39488342-aa1a-4dba-bd5f-556da5081904" path="/var/lib/kubelet/pods/39488342-aa1a-4dba-bd5f-556da5081904/volumes" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.869499 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.898210 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.909555 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: E0313 14:30:03.910102 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerName="collect-profiles" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910126 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerName="collect-profiles" Mar 13 14:30:03 crc kubenswrapper[4907]: E0313 14:30:03.910174 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910367 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" containerName="nova-scheduler-scheduler" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.910389 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" containerName="collect-profiles" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.911079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.913194 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.919640 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.973143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.973355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:03 crc kubenswrapper[4907]: I0313 14:30:03.973562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.075803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.075910 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.076000 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.081046 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.081417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.096679 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"nova-scheduler-0\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.218945 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.229298 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.279340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") pod \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\" (UID: \"e7993f31-1072-4c83-a3ac-16dca5eb04a3\") " Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.283927 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz" (OuterVolumeSpecName: "kube-api-access-2kxlz") pod "e7993f31-1072-4c83-a3ac-16dca5eb04a3" (UID: "e7993f31-1072-4c83-a3ac-16dca5eb04a3"). InnerVolumeSpecName "kube-api-access-2kxlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.382252 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kxlz\" (UniqueName: \"kubernetes.io/projected/e7993f31-1072-4c83-a3ac-16dca5eb04a3-kube-api-access-2kxlz\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.665828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:04 crc kubenswrapper[4907]: W0313 14:30:04.669687 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4050d00b_0256_45c5_9dc4_0ab46956405d.slice/crio-c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc WatchSource:0}: Error finding container c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc: Status 404 returned error can't find the container with id c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.782360 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.782359 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556870-8mlkz" event={"ID":"e7993f31-1072-4c83-a3ac-16dca5eb04a3","Type":"ContainerDied","Data":"0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da"} Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.782527 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fff58df59ba6aeb1bad1085410d3a35e667921192b7b80aca529cb3350038da" Mar 13 14:30:04 crc kubenswrapper[4907]: I0313 14:30:04.785221 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerStarted","Data":"c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc"} Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.309096 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.317931 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556864-768cg"] Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.796046 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32f92b7e-159a-4824-93e7-ea6d49961db8" path="/var/lib/kubelet/pods/32f92b7e-159a-4824-93e7-ea6d49961db8/volumes" Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.796827 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0304716-7acd-4a2c-a018-73fc4bcbad51" path="/var/lib/kubelet/pods/c0304716-7acd-4a2c-a018-73fc4bcbad51/volumes" Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.797348 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerStarted","Data":"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56"} Mar 13 14:30:05 crc kubenswrapper[4907]: I0313 14:30:05.821533 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.821514209 podStartE2EDuration="2.821514209s" podCreationTimestamp="2026-03-13 14:30:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 14:30:05.813397225 +0000 UTC m=+1504.713184914" watchObservedRunningTime="2026-03-13 14:30:05.821514209 +0000 UTC m=+1504.721301898" Mar 13 14:30:08 crc kubenswrapper[4907]: I0313 14:30:08.064113 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:30:08 crc kubenswrapper[4907]: I0313 14:30:08.064402 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 14:30:09 crc kubenswrapper[4907]: I0313 14:30:09.078066 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:09 crc kubenswrapper[4907]: I0313 14:30:09.078154 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.211:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:09 crc kubenswrapper[4907]: I0313 14:30:09.230074 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 14:30:12 crc kubenswrapper[4907]: I0313 14:30:12.159551 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:30:12 crc kubenswrapper[4907]: I0313 14:30:12.160632 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 14:30:13 crc kubenswrapper[4907]: I0313 14:30:13.174061 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:13 crc kubenswrapper[4907]: I0313 14:30:13.174100 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.214:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:14 crc kubenswrapper[4907]: I0313 14:30:14.229519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 14:30:14 crc kubenswrapper[4907]: I0313 14:30:14.255946 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 14:30:14 crc kubenswrapper[4907]: I0313 14:30:14.916492 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 14:30:15 crc kubenswrapper[4907]: I0313 14:30:15.809706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 14:30:16 crc kubenswrapper[4907]: I0313 14:30:16.063585 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:30:16 crc kubenswrapper[4907]: I0313 14:30:16.063640 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.073950 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.075509 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.098149 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:30:18 crc kubenswrapper[4907]: I0313 14:30:18.927784 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 14:30:20 crc kubenswrapper[4907]: I0313 14:30:20.159037 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:30:20 crc kubenswrapper[4907]: I0313 14:30:20.159533 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 14:30:21 crc kubenswrapper[4907]: I0313 14:30:21.726971 4907 scope.go:117] "RemoveContainer" containerID="a53552e35726d3cc22de750fc04d8a15ada7be07f868ed527291e30242034c7d" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.165168 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.165277 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.170567 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 14:30:22 crc kubenswrapper[4907]: I0313 14:30:22.173425 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.630126 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.647704 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-r6vk2"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.679011 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:39 crc kubenswrapper[4907]: E0313 14:30:39.679496 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerName="oc" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.679515 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerName="oc" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.679759 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" containerName="oc" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.680512 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.682336 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.709267 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.751296 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.751696 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" containerID="cri-o://3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88" gracePeriod=2 Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.771079 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.778652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.778790 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.816340 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b4fd44-4a11-4521-bfd0-16c5436368fa" path="/var/lib/kubelet/pods/d1b4fd44-4a11-4521-bfd0-16c5436368fa/volumes" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.883173 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.883346 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.884371 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.961040 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:30:39 crc kubenswrapper[4907]: I0313 14:30:39.965724 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"root-account-create-update-glzbl\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.009017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.012502 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9c64-account-create-update-v65tv"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.040379 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.041851 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" containerID="cri-o://1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6" gracePeriod=300 Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.088204 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.088269 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.088348 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:40.588317102 +0000 UTC m=+1539.488104791 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.291960 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.310937 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.310979 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.311752 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerName="openstackclient" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.312611 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.366629 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.366783 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.408447 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.413035 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.424146 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.445407 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.474560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.474634 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.475393 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.497814 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.529226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"placement-9c64-account-create-update-5vvgg\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.550451 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.552086 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.555374 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.576780 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.576853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.577379 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.585927 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.596965 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.631936 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-47e2-account-create-update-r4ccm"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.664700 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685588 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685689 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.685744 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.704831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704926 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704941 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704952 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.704993 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:41.704981383 +0000 UTC m=+1540.604769072 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.709661 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.719716 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" containerID="cri-o://432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad" gracePeriod=300 Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.746312 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.789395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.789605 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.791617 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:40 crc kubenswrapper[4907]: E0313 14:30:40.791664 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:41.291648284 +0000 UTC m=+1540.191435973 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.792203 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.807602 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"cinder-47e2-account-create-update-wvjr7\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.808097 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.810438 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.824209 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" probeResult="failure" output="command timed out" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.824986 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.828917 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" probeResult="failure" output="command timed out" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.837574 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"neutron-b69e-account-create-update-96wq7\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.849304 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b69e-account-create-update-ptpcx"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.871951 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.892000 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.892169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.914711 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.959118 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.959754 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" containerID="cri-o://f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" gracePeriod=300 Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.988746 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.995724 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.996011 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:40 crc kubenswrapper[4907]: I0313 14:30:40.996985 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.030935 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-17a8-account-create-update-xv8sm"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.047283 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"glance-17a8-account-create-update-nj9pw\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.074070 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.074293 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" containerID="cri-o://e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.074663 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" containerID="cri-o://59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.102908 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.124587 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.175401 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.190782 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7q999"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.223312 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.253349 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" containerID="cri-o://46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" gracePeriod=300 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.270452 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jt88g"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.308115 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.315646 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.315770 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:42.3157553 +0000 UTC m=+1541.215542989 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.342826 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-jjqr9"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.377264 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.396939 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-22bc-account-create-update-r969t"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398634 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/ovsdbserver-sb/0.log" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398674 4907 generic.go:334] "Generic (PLEG): container finished" podID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerID="1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6" exitCode=2 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398689 4907 generic.go:334] "Generic (PLEG): container finished" podID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerID="432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad" exitCode=143 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.398935 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerDied","Data":"1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.399115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerDied","Data":"432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.419134 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420414 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_26827945-75f0-4867-ba04-31ff6428e06a/ovn-northd/0.log" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420445 4907 generic.go:334] "Generic (PLEG): container finished" podID="26827945-75f0-4867-ba04-31ff6428e06a" containerID="59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508" exitCode=2 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420460 4907 generic.go:334] "Generic (PLEG): container finished" podID="26827945-75f0-4867-ba04-31ff6428e06a" containerID="e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b" exitCode=143 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420515 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerDied","Data":"59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.420535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerDied","Data":"e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.426597 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_056dd756-0d7b-471b-9929-f622d05ad606/ovsdbserver-nb/0.log" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.426637 4907 generic.go:334] "Generic (PLEG): container finished" podID="056dd756-0d7b-471b-9929-f622d05ad606" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" exitCode=2 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.426666 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerDied","Data":"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb"} Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.453484 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.464988 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.493788 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-rz6qh"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.548687 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.549207 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" containerID="cri-o://890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1" gracePeriod=10 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.621022 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.656032 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.656283 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-nhnkq" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" containerID="cri-o://62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.686872 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.722163 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-9fxfh"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752786 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752819 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752828 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752838 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.752900 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:43.752872436 +0000 UTC m=+1542.652660125 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.778498 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f148-account-create-update-wgw4b"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.816472 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0983a1bd-a83c-4658-a405-09b8ab0d0002" path="/var/lib/kubelet/pods/0983a1bd-a83c-4658-a405-09b8ab0d0002/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.817340 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32b531c6-64c1-4137-b82e-3c10789325e6" path="/var/lib/kubelet/pods/32b531c6-64c1-4137-b82e-3c10789325e6/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.817852 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3db3f0b6-9cae-40fa-b54d-3ed06c568c9c" path="/var/lib/kubelet/pods/3db3f0b6-9cae-40fa-b54d-3ed06c568c9c/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.819203 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fe8116b-9e68-41e2-aad2-cdd7d5e51c83" path="/var/lib/kubelet/pods/3fe8116b-9e68-41e2-aad2-cdd7d5e51c83/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.820256 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6" path="/var/lib/kubelet/pods/4eeb2fb2-22fb-4ea8-a39b-5f959877cfe6/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.820797 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="594b61fa-12b7-47b6-8af3-17f024e0d54d" path="/var/lib/kubelet/pods/594b61fa-12b7-47b6-8af3-17f024e0d54d/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.821606 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85913727-bbcd-4343-9faf-a75f40b42dc8" path="/var/lib/kubelet/pods/85913727-bbcd-4343-9faf-a75f40b42dc8/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.824417 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="870ec55c-fc6f-4b16-8c49-a3cda0d0d010" path="/var/lib/kubelet/pods/870ec55c-fc6f-4b16-8c49-a3cda0d0d010/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.825364 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b144897c-8640-4c5c-afa1-68b44dfd4f98" path="/var/lib/kubelet/pods/b144897c-8640-4c5c-afa1-68b44dfd4f98/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.826212 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb6c089-0112-43f4-8731-c2d68932795a" path="/var/lib/kubelet/pods/cbb6c089-0112-43f4-8731-c2d68932795a/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.827774 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2585497-6837-4d17-8d51-c3d7879fdb46" path="/var/lib/kubelet/pods/e2585497-6837-4d17-8d51-c3d7879fdb46/volumes" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.828976 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4v4ps"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.845018 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.854267 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" cmd=["/usr/bin/pidof","ovsdb-server"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.863042 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" cmd=["/usr/bin/pidof","ovsdb-server"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.869780 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" cmd=["/usr/bin/pidof","ovsdb-server"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.869845 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.874947 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.875174 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" containerID="cri-o://d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.875516 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" containerID="cri-o://94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.901306 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.918671 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.949605 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:41 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: if [ -n "neutron" ]; then Mar 13 14:30:41 crc kubenswrapper[4907]: GRANT_DATABASE="neutron" Mar 13 14:30:41 crc kubenswrapper[4907]: else Mar 13 14:30:41 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:41 crc kubenswrapper[4907]: fi Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:41 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:41 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:41 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:41 crc kubenswrapper[4907]: # support updates Mar 13 14:30:41 crc kubenswrapper[4907]: Mar 13 14:30:41 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:41 crc kubenswrapper[4907]: E0313 14:30:41.951751 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"neutron-db-secret\\\" not found\"" pod="openstack/neutron-b69e-account-create-update-96wq7" podUID="e8f6894c-b714-431e-912d-8b8de1a3914b" Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.956919 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.957718 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bf587b8d9-pvvbz" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" containerID="cri-o://8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.958442 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bf587b8d9-pvvbz" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" containerID="cri-o://3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.978210 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f6fe-account-create-update-56786"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.996294 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.996616 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-85dc486cc4-fwrm6" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" containerID="cri-o://c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" gracePeriod=30 Mar 13 14:30:41 crc kubenswrapper[4907]: I0313 14:30:41.999216 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-85dc486cc4-fwrm6" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" containerID="cri-o://6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.020958 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.021472 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" containerID="cri-o://2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.021985 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" containerID="cri-o://a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.042576 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043313 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" containerID="cri-o://8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043423 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" containerID="cri-o://dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043461 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" containerID="cri-o://7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043503 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" containerID="cri-o://2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043531 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" containerID="cri-o://b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043560 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" containerID="cri-o://c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043589 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" containerID="cri-o://f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043619 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" containerID="cri-o://9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043734 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" containerID="cri-o://8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043418 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" containerID="cri-o://c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043889 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" containerID="cri-o://5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043832 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" containerID="cri-o://1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043848 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" containerID="cri-o://0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043857 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" containerID="cri-o://c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.043867 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" containerID="cri-o://6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.057241 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.067819 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.071449 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:42.57142518 +0000 UTC m=+1541.471212869 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.087988 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-6g25t"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.138788 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:30:42 crc kubenswrapper[4907]: W0313 14:30:42.146245 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a4c64a3_7378_4d8d_9cf3_17583c6a90c0.slice/crio-a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098 WatchSource:0}: Error finding container a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098: Status 404 returned error can't find the container with id a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.158299 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2pvp9"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.176445 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: if [ -n "placement" ]; then Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="placement" Mar 13 14:30:42 crc kubenswrapper[4907]: else Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:42 crc kubenswrapper[4907]: fi Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:42 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:42 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:42 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:42 crc kubenswrapper[4907]: # support updates Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.177762 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"placement-db-secret\\\" not found\"" pod="openstack/placement-9c64-account-create-update-5vvgg" podUID="4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.182868 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.194275 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.194540 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" containerID="cri-o://676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.194669 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" containerID="cri-o://c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.201131 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.201335 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" containerID="cri-o://70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.201453 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" containerID="cri-o://479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.208949 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.226621 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-x9b2b"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.259793 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.260023 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" containerID="cri-o://eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.260456 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" containerID="cri-o://0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.280288 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.296975 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.329507 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-cnnxs"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.377911 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.378145 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" containerID="cri-o://3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.378565 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" containerID="cri-o://7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.385360 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.385439 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:44.385419186 +0000 UTC m=+1543.285206875 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.389312 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.401983 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.402203 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" containerID="cri-o://b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.402561 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" containerID="cri-o://be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.418165 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.497182 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.502674 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" containerID="cri-o://f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.502807 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" containerID="cri-o://22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.513561 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_056dd756-0d7b-471b-9929-f622d05ad606/ovsdbserver-nb/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.513668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.524308 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.524524 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" containerID="cri-o://c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.524676 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" containerID="cri-o://ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.525019 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_26827945-75f0-4867-ba04-31ff6428e06a/ovn-northd/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.525089 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.536529 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.538675 4907 generic.go:334] "Generic (PLEG): container finished" podID="4251ae36-90ad-41ea-915e-862df60f5c07" containerID="70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.538732 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerDied","Data":"70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.550629 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.550683 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerDied","Data":"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.566555 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.581156 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nhnkq_37978cc1-fcba-4032-a8b1-6632b61692ff/openstack-network-exporter/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.583616 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.585184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-5vvgg" event={"ID":"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0","Type":"ContainerStarted","Data":"a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.592218 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-g96mx"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.595244 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: if [ -n "placement" ]; then Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="placement" Mar 13 14:30:42 crc kubenswrapper[4907]: else Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:42 crc kubenswrapper[4907]: fi Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:42 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:42 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:42 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:42 crc kubenswrapper[4907]: # support updates Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.598283 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"placement-db-secret\\\" not found\"" pod="openstack/placement-9c64-account-create-update-5vvgg" podUID="4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.603867 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-96wq7" event={"ID":"e8f6894c-b714-431e-912d-8b8de1a3914b","Type":"ContainerStarted","Data":"d38b680247762c5b2510a8776b9df1c80f0da04bd9de35e5d2db2452cf7446f0"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.607151 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/ovsdbserver-sb/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.607234 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.607946 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608164 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608278 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608318 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608337 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608401 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608418 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608501 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608565 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608659 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608680 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") pod \"26827945-75f0-4867-ba04-31ff6428e06a\" (UID: \"26827945-75f0-4867-ba04-31ff6428e06a\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.608713 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") pod \"056dd756-0d7b-471b-9929-f622d05ad606\" (UID: \"056dd756-0d7b-471b-9929-f622d05ad606\") " Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.609181 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.609221 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:43.609209046 +0000 UTC m=+1542.508996735 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.612368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config" (OuterVolumeSpecName: "config") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.615194 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.616822 4907 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 14:30:42 crc kubenswrapper[4907]: + source /usr/local/bin/container-scripts/functions Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNBridge=br-int Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNRemote=tcp:localhost:6642 Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNEncapType=geneve Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNAvailabilityZones= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ EnableChassisAsGateway=true Mar 13 14:30:42 crc kubenswrapper[4907]: ++ PhysicalNetworks= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNHostName= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 14:30:42 crc kubenswrapper[4907]: ++ ovs_dir=/var/lib/openvswitch Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 14:30:42 crc kubenswrapper[4907]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + sleep 0.5 Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + cleanup_ovsdb_server_semaphore Mar 13 14:30:42 crc kubenswrapper[4907]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-vp6f5" message=< Mar 13 14:30:42 crc kubenswrapper[4907]: Exiting ovsdb-server (5) ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 14:30:42 crc kubenswrapper[4907]: + source /usr/local/bin/container-scripts/functions Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNBridge=br-int Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNRemote=tcp:localhost:6642 Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNEncapType=geneve Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNAvailabilityZones= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ EnableChassisAsGateway=true Mar 13 14:30:42 crc kubenswrapper[4907]: ++ PhysicalNetworks= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNHostName= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 14:30:42 crc kubenswrapper[4907]: ++ ovs_dir=/var/lib/openvswitch Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 14:30:42 crc kubenswrapper[4907]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + sleep 0.5 Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + cleanup_ovsdb_server_semaphore Mar 13 14:30:42 crc kubenswrapper[4907]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.616895 4907 kuberuntime_container.go:691] "PreStop hook failed" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 14:30:42 crc kubenswrapper[4907]: + source /usr/local/bin/container-scripts/functions Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNBridge=br-int Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNRemote=tcp:localhost:6642 Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNEncapType=geneve Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNAvailabilityZones= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ EnableChassisAsGateway=true Mar 13 14:30:42 crc kubenswrapper[4907]: ++ PhysicalNetworks= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ OVNHostName= Mar 13 14:30:42 crc kubenswrapper[4907]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 14:30:42 crc kubenswrapper[4907]: ++ ovs_dir=/var/lib/openvswitch Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 14:30:42 crc kubenswrapper[4907]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 14:30:42 crc kubenswrapper[4907]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + sleep 0.5 Mar 13 14:30:42 crc kubenswrapper[4907]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 14:30:42 crc kubenswrapper[4907]: + cleanup_ovsdb_server_semaphore Mar 13 14:30:42 crc kubenswrapper[4907]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 14:30:42 crc kubenswrapper[4907]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" containerID="cri-o://6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.616938 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" containerID="cri-o://6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.617321 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.621217 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.622003 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts" (OuterVolumeSpecName: "scripts") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.622927 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts" (OuterVolumeSpecName: "scripts") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.636384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.638020 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-htpvb"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.638176 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config" (OuterVolumeSpecName: "config") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.649121 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:42 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: if [ -n "neutron" ]; then Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="neutron" Mar 13 14:30:42 crc kubenswrapper[4907]: else Mar 13 14:30:42 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:42 crc kubenswrapper[4907]: fi Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:42 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:42 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:42 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:42 crc kubenswrapper[4907]: # support updates Mar 13 14:30:42 crc kubenswrapper[4907]: Mar 13 14:30:42 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.654152 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"neutron-db-secret\\\" not found\"" pod="openstack/neutron-b69e-account-create-update-96wq7" podUID="e8f6894c-b714-431e-912d-8b8de1a3914b" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.655423 4907 generic.go:334] "Generic (PLEG): container finished" podID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerID="2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.655542 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerDied","Data":"2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.660278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.662074 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" probeResult="failure" output=< Mar 13 14:30:42 crc kubenswrapper[4907]: 2026-03-13T14:30:42Z|00001|jsonrpc|WARN|unix:/var/run/openvswitch/ovs-vswitchd.11.ctl: receive error: Connection reset by peer Mar 13 14:30:42 crc kubenswrapper[4907]: 2026-03-13T14:30:42Z|00002|unixctl|WARN|error communicating with unix:/var/run/openvswitch/ovs-vswitchd.11.ctl: Connection reset by peer Mar 13 14:30:42 crc kubenswrapper[4907]: ovs-appctl: /var/run/openvswitch/ovs-vswitchd.11.ctl: transaction error (Connection reset by peer) Mar 13 14:30:42 crc kubenswrapper[4907]: ERROR - Failed retrieving ofproto/list from ovs-vswitchd Mar 13 14:30:42 crc kubenswrapper[4907]: > Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.667297 4907 generic.go:334] "Generic (PLEG): container finished" podID="516d1499-3461-4af5-b426-9ae4a711a468" containerID="783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4" exitCode=1 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.668255 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerDied","Data":"783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.669833 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerStarted","Data":"f6d2f722265ae72bf48903452fd81587c07136e043b3e098f648b2d79cb3fd59"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.669770 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j" (OuterVolumeSpecName: "kube-api-access-lgj9j") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "kube-api-access-lgj9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.673147 4907 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/root-account-create-update-glzbl" secret="" err="secret \"galera-openstack-cell1-dockercfg-szj6q\" not found" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.673277 4907 scope.go:117] "RemoveContainer" containerID="783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.676934 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.690303 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p" (OuterVolumeSpecName: "kube-api-access-ckz4p") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "kube-api-access-ckz4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.691832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713453 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713490 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713550 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713583 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713643 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713663 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713680 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713735 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713762 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713785 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713804 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713838 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713915 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713937 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.713956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.714001 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") pod \"37978cc1-fcba-4032-a8b1-6632b61692ff\" (UID: \"37978cc1-fcba-4032-a8b1-6632b61692ff\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.714029 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") pod \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\" (UID: \"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.714512 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.715217 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/26827945-75f0-4867-ba04-31ff6428e06a-ovn-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.725326 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726469 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgj9j\" (UniqueName: \"kubernetes.io/projected/26827945-75f0-4867-ba04-31ff6428e06a-kube-api-access-lgj9j\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726537 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/056dd756-0d7b-471b-9929-f622d05ad606-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726609 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.727407 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/056dd756-0d7b-471b-9929-f622d05ad606-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.727472 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26827945-75f0-4867-ba04-31ff6428e06a-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.727539 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckz4p\" (UniqueName: \"kubernetes.io/projected/056dd756-0d7b-471b-9929-f622d05ad606-kube-api-access-ckz4p\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.721764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config" (OuterVolumeSpecName: "config") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.722341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config" (OuterVolumeSpecName: "config") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.723998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726197 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.726225 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.716182 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" containerID="cri-o://d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" gracePeriod=604800 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.716360 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_26827945-75f0-4867-ba04-31ff6428e06a/ovn-northd/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.721036 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b501-account-create-update-dzlqr"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.731207 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.736586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"26827945-75f0-4867-ba04-31ff6428e06a","Type":"ContainerDied","Data":"59513110b8b488ed882794dedcca28bd1c64f5a6610bf84e8eaeafbea7ce6daa"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.736797 4907 scope.go:117] "RemoveContainer" containerID="59dc65eade55429a5d17c956c987a8209b624ac898dd355bca1d93bdf510b508" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.728608 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.730662 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.740620 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts" (OuterVolumeSpecName: "scripts") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.752595 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s" (OuterVolumeSpecName: "kube-api-access-pxn9s") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "kube-api-access-pxn9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.753580 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" containerID="cri-o://265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.754908 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb" (OuterVolumeSpecName: "kube-api-access-g2kbb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "kube-api-access-g2kbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.756359 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5" (OuterVolumeSpecName: "kube-api-access-ndzf5") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "kube-api-access-ndzf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.756677 4907 generic.go:334] "Generic (PLEG): container finished" podID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerID="eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.756766 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerDied","Data":"eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.760381 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-sgpkl"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.768764 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.779234 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.788032 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795181 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795220 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795232 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795240 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795248 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795256 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795264 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795273 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795281 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795289 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795336 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795381 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795394 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795405 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795418 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795430 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795453 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795461 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795629 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.795849 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" containerID="cri-o://6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.796257 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" containerID="cri-o://58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.796368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.813771 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_056dd756-0d7b-471b-9929-f622d05ad606/ovsdbserver-nb/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.813822 4907 generic.go:334] "Generic (PLEG): container finished" podID="056dd756-0d7b-471b-9929-f622d05ad606" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" exitCode=143 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.813946 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.814582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerDied","Data":"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.814639 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"056dd756-0d7b-471b-9929-f622d05ad606","Type":"ContainerDied","Data":"8f94280adbcbea9567f862d27736043270ac488194f6b0213cdb79c3aa5a9dee"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829773 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829801 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829810 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovn-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829819 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37978cc1-fcba-4032-a8b1-6632b61692ff-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829827 4907 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/37978cc1-fcba-4032-a8b1-6632b61692ff-ovs-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829835 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxn9s\" (UniqueName: \"kubernetes.io/projected/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-kube-api-access-pxn9s\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829844 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829853 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829871 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829938 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829949 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829958 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ndzf5\" (UniqueName: \"kubernetes.io/projected/37978cc1-fcba-4032-a8b1-6632b61692ff-kube-api-access-ndzf5\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.829967 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2kbb\" (UniqueName: \"kubernetes.io/projected/7da13722-732a-4b65-a894-fe4612f30d75-kube-api-access-g2kbb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.830473 4907 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Mar 13 14:30:42 crc kubenswrapper[4907]: E0313 14:30:42.830559 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts podName:516d1499-3461-4af5-b426-9ae4a711a468 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:43.330540211 +0000 UTC m=+1542.230327900 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts") pod "root-account-create-update-glzbl" (UID: "516d1499-3461-4af5-b426-9ae4a711a468") : configmap "openstack-cell1-scripts" not found Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.843009 4907 scope.go:117] "RemoveContainer" containerID="e23ef8c27779c528c50bff3fd731469de16d17d8b86a266ab0618151d6dd930b" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852429 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nhnkq_37978cc1-fcba-4032-a8b1-6632b61692ff/openstack-network-exporter/0.log" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852476 4907 generic.go:334] "Generic (PLEG): container finished" podID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerID="62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a" exitCode=2 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852595 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nhnkq" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.852910 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nhnkq" event={"ID":"37978cc1-fcba-4032-a8b1-6632b61692ff","Type":"ContainerDied","Data":"62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.855166 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.856710 4907 generic.go:334] "Generic (PLEG): container finished" podID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" containerID="3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88" exitCode=137 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.859437 4907 generic.go:334] "Generic (PLEG): container finished" podID="7da13722-732a-4b65-a894-fe4612f30d75" containerID="890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1" exitCode=0 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.859466 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" event={"ID":"7da13722-732a-4b65-a894-fe4612f30d75","Type":"ContainerDied","Data":"890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1"} Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.859574 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fdb8f6449-nb4km" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.862413 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.869661 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vlp6r"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.874198 4907 scope.go:117] "RemoveContainer" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.877963 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.890107 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.904746 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.916688 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.923424 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z92h5"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.927051 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.928661 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ldbt9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.933893 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.933958 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") pod \"7da13722-732a-4b65-a894-fe4612f30d75\" (UID: \"7da13722-732a-4b65-a894-fe4612f30d75\") " Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.934496 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:42 crc kubenswrapper[4907]: W0313 14:30:42.935387 4907 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/7da13722-732a-4b65-a894-fe4612f30d75/volumes/kubernetes.io~configmap/ovsdbserver-nb Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.935409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.936679 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vbcql"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.943201 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.943387 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" containerID="cri-o://5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.949717 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.949870 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" containerID="cri-o://77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" gracePeriod=30 Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.956774 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.968362 4907 scope.go:117] "RemoveContainer" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.972705 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-xw7n9"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.981246 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:30:42 crc kubenswrapper[4907]: I0313 14:30:42.981646 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7" gracePeriod=30 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:42.993609 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:42.995234 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "056dd756-0d7b-471b-9929-f622d05ad606" (UID: "056dd756-0d7b-471b-9929-f622d05ad606"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.000864 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.001587 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.039164 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.039428 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.039438 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/056dd756-0d7b-471b-9929-f622d05ad606-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.045190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.046040 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:43 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: if [ -n "cinder" ]; then Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="cinder" Mar 13 14:30:43 crc kubenswrapper[4907]: else Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:43 crc kubenswrapper[4907]: fi Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:43 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:43 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:43 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:43 crc kubenswrapper[4907]: # support updates Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.048426 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"cinder-db-secret\\\" not found\"" pod="openstack/cinder-47e2-account-create-update-wvjr7" podUID="82dec019-ba60-43f5-845b-322a6bcca225" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.105781 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.140918 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.140960 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.143941 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" containerID="cri-o://af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" gracePeriod=30 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.159400 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config" (OuterVolumeSpecName: "config") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.163169 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.174249 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.174502 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" containerID="cri-o://8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" gracePeriod=30 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.202302 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.202794 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" containerID="cri-o://c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4" gracePeriod=604800 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.213613 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.220922 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37978cc1-fcba-4032-a8b1-6632b61692ff" (UID: "37978cc1-fcba-4032-a8b1-6632b61692ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.226327 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246207 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246242 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246251 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246260 4907 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.246268 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37978cc1-fcba-4032-a8b1-6632b61692ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.254639 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7da13722-732a-4b65-a894-fe4612f30d75" (UID: "7da13722-732a-4b65-a894-fe4612f30d75"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.266724 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.268091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "26827945-75f0-4867-ba04-31ff6428e06a" (UID: "26827945-75f0-4867-ba04-31ff6428e06a"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.286093 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" (UID: "c2938e39-f1d8-4cdc-a32c-5d57b8f2034f"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.338254 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.340219 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348110 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7da13722-732a-4b65-a894-fe4612f30d75-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348145 4907 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348159 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.348197 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/26827945-75f0-4867-ba04-31ff6428e06a-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.348283 4907 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.348342 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts podName:516d1499-3461-4af5-b426-9ae4a711a468 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:44.348325611 +0000 UTC m=+1543.248113300 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts") pod "root-account-create-update-glzbl" (UID: "516d1499-3461-4af5-b426-9ae4a711a468") : configmap "openstack-cell1-scripts" not found Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.357042 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.358635 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.358705 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.419851 4907 scope.go:117] "RemoveContainer" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.420565 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb\": container with ID starting with f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb not found: ID does not exist" containerID="f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.420617 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb"} err="failed to get container status \"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb\": rpc error: code = NotFound desc = could not find container \"f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb\": container with ID starting with f294f1dfa23d177ca2e8f5511b6bd495bd3146c4baee6d3b95ccd81e1b4496fb not found: ID does not exist" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.420650 4907 scope.go:117] "RemoveContainer" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.422089 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab\": container with ID starting with 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab not found: ID does not exist" containerID="46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.422108 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab"} err="failed to get container status \"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab\": rpc error: code = NotFound desc = could not find container \"46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab\": container with ID starting with 46f9fcec5ce705120f46ecb4deec4665885423d12b23cabce7f929fa5941d7ab not found: ID does not exist" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.422124 4907 scope.go:117] "RemoveContainer" containerID="62edff4734da4ab22bc7673f1d2c0e9a9db082a3ca22b12fadc41486071fe66a" Mar 13 14:30:43 crc kubenswrapper[4907]: W0313 14:30:43.426227 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc15f080_f90d_4e97_a866_d5edfd662bb7.slice/crio-bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63 WatchSource:0}: Error finding container bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63: Status 404 returned error can't find the container with id bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63 Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.428678 4907 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 14:30:43 crc kubenswrapper[4907]: container &Container{Name:mariadb-account-create-update,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:4caef2b55e01b9a7ee88a22bc69db1893521a91d95c7ad4c8e593f14f17a5f95,Command:[/bin/sh -c #!/bin/bash Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_REMOTE_HOST="" source /var/lib/operator-scripts/mysql_root_auth.sh Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: export DatabasePassword=${DatabasePassword:?"Please specify a DatabasePassword variable."} Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: MYSQL_CMD="mysql -h -u root -P 3306" Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: if [ -n "glance" ]; then Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="glance" Mar 13 14:30:43 crc kubenswrapper[4907]: else Mar 13 14:30:43 crc kubenswrapper[4907]: GRANT_DATABASE="*" Mar 13 14:30:43 crc kubenswrapper[4907]: fi Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: # going for maximum compatibility here: Mar 13 14:30:43 crc kubenswrapper[4907]: # 1. MySQL 8 no longer allows implicit create user when GRANT is used Mar 13 14:30:43 crc kubenswrapper[4907]: # 2. MariaDB has "CREATE OR REPLACE", but MySQL does not Mar 13 14:30:43 crc kubenswrapper[4907]: # 3. create user with CREATE but then do all password and TLS with ALTER to Mar 13 14:30:43 crc kubenswrapper[4907]: # support updates Mar 13 14:30:43 crc kubenswrapper[4907]: Mar 13 14:30:43 crc kubenswrapper[4907]: $MYSQL_CMD < logger="UnhandledError" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.438529 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-create-update\" with CreateContainerConfigError: \"secret \\\"glance-db-secret\\\" not found\"" pod="openstack/glance-17a8-account-create-update-nj9pw" podUID="bc15f080-f90d-4e97-a866-d5edfd662bb7" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.450499 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.459374 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.459982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.476210 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.483858 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.486550 4907 scope.go:117] "RemoveContainer" containerID="890693bd8cb16d526cbd4e7c94b7a9b8dcd67a1bc48dbb881c81f86b4e9727d1" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.491738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.511276 4907 scope.go:117] "RemoveContainer" containerID="9c33ada0f7c03cde714537a63faadeb180f6ac07dd15e3f4ce8c7087574733bb" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.572109 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.584571 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-nhnkq"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.591534 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.601067 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fdb8f6449-nb4km"] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671635 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671678 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671754 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671804 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671916 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671946 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.671972 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.672000 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") pod \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\" (UID: \"a055ce7f-0538-4e5e-938a-eaf6d0aa2dec\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.672061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.672083 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") pod \"1a1494c5-e2d6-4d29-b161-97b720742d8d\" (UID: \"1a1494c5-e2d6-4d29-b161-97b720742d8d\") " Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.672535 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.672591 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:45.672576966 +0000 UTC m=+1544.572364655 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.675311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.676612 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k" (OuterVolumeSpecName: "kube-api-access-fk82k") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "kube-api-access-fk82k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.677385 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.682180 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.682545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6" (OuterVolumeSpecName: "kube-api-access-7gbp6") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "kube-api-access-7gbp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.717023 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.728365 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.756635 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774389 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774418 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774465 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774479 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:43 crc kubenswrapper[4907]: E0313 14:30:43.774533 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:47.774510309 +0000 UTC m=+1546.674297998 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775008 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk82k\" (UniqueName: \"kubernetes.io/projected/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-kube-api-access-fk82k\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775027 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775038 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1a1494c5-e2d6-4d29-b161-97b720742d8d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775048 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gbp6\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-kube-api-access-7gbp6\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775291 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775305 4907 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1a1494c5-e2d6-4d29-b161-97b720742d8d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.775313 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.786667 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.790831 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.797262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data" (OuterVolumeSpecName: "config-data") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.798540 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a1494c5-e2d6-4d29-b161-97b720742d8d" (UID: "1a1494c5-e2d6-4d29-b161-97b720742d8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.802611 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02262347-eaec-462e-a3ef-58c670885d99" path="/var/lib/kubelet/pods/02262347-eaec-462e-a3ef-58c670885d99/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.803438 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="056dd756-0d7b-471b-9929-f622d05ad606" path="/var/lib/kubelet/pods/056dd756-0d7b-471b-9929-f622d05ad606/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.804145 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b701a42-f649-4657-9250-49188b10d3b9" path="/var/lib/kubelet/pods/1b701a42-f649-4657-9250-49188b10d3b9/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.805091 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="251ab76e-b9df-406f-9b07-c7fa6e227feb" path="/var/lib/kubelet/pods/251ab76e-b9df-406f-9b07-c7fa6e227feb/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.805644 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26827945-75f0-4867-ba04-31ff6428e06a" path="/var/lib/kubelet/pods/26827945-75f0-4867-ba04-31ff6428e06a/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.806264 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bab6ed6-5335-40cc-bdff-3d7461876fb8" path="/var/lib/kubelet/pods/2bab6ed6-5335-40cc-bdff-3d7461876fb8/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.807272 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" path="/var/lib/kubelet/pods/37978cc1-fcba-4032-a8b1-6632b61692ff/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.807846 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f2b7163-b5d4-4575-a72c-e6ad1b3137df" path="/var/lib/kubelet/pods/3f2b7163-b5d4-4575-a72c-e6ad1b3137df/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.808463 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7d2639-4229-4286-be5c-d8b15ed91d17" path="/var/lib/kubelet/pods/4d7d2639-4229-4286-be5c-d8b15ed91d17/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.809365 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7ea7a2-ab25-4cb4-8749-10bc129cbd22" path="/var/lib/kubelet/pods/4d7ea7a2-ab25-4cb4-8749-10bc129cbd22/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.810613 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dcaf7c6-b4dd-4312-b68b-a833a827a2c2" path="/var/lib/kubelet/pods/5dcaf7c6-b4dd-4312-b68b-a833a827a2c2/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.811186 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65556420-dc7f-49c8-af59-abafebbb4674" path="/var/lib/kubelet/pods/65556420-dc7f-49c8-af59-abafebbb4674/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.811641 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af00fde-8a92-4530-8fa8-9df7ff787c93" path="/var/lib/kubelet/pods/7af00fde-8a92-4530-8fa8-9df7ff787c93/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.813942 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7da13722-732a-4b65-a894-fe4612f30d75" path="/var/lib/kubelet/pods/7da13722-732a-4b65-a894-fe4612f30d75/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.814461 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8901c771-b45a-4147-9f96-cef784165281" path="/var/lib/kubelet/pods/8901c771-b45a-4147-9f96-cef784165281/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.815232 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9118baf0-df6a-41e5-98e5-3e4d8edfbaaa" path="/var/lib/kubelet/pods/9118baf0-df6a-41e5-98e5-3e4d8edfbaaa/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.816444 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98748d4d-f934-49ed-91d9-7a9e467b2753" path="/var/lib/kubelet/pods/98748d4d-f934-49ed-91d9-7a9e467b2753/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.817005 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a89525da-db62-49c2-b548-bab4c974dfdd" path="/var/lib/kubelet/pods/a89525da-db62-49c2-b548-bab4c974dfdd/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.817517 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1068e89-10df-4d4c-a213-d9e9643fab38" path="/var/lib/kubelet/pods/e1068e89-10df-4d4c-a213-d9e9643fab38/volumes" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.828808 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" (UID: "a055ce7f-0538-4e5e-938a-eaf6d0aa2dec"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.871723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-nj9pw" event={"ID":"bc15f080-f90d-4e97-a866-d5edfd662bb7","Type":"ContainerStarted","Data":"bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.875601 4907 generic.go:334] "Generic (PLEG): container finished" podID="c1b70392-1240-40d9-8128-e7abe29c8398" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.875723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876357 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.876474 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877025 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877100 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877144 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") pod \"3793f53a-ca36-4b27-8444-d6dbfd860424\" (UID: \"3793f53a-ca36-4b27-8444-d6dbfd860424\") " Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.877964 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878000 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878010 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878021 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878030 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3793f53a-ca36-4b27-8444-d6dbfd860424-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.878038 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1494c5-e2d6-4d29-b161-97b720742d8d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.882049 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.882290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts" (OuterVolumeSpecName: "scripts") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.884471 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.884520 4907 scope.go:117] "RemoveContainer" containerID="3e056a670aa7307d9d211e2123e0cead9b6cd31175ff09e1feddb508c47f1a88" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.888227 4907 generic.go:334] "Generic (PLEG): container finished" podID="62983e04-f043-4ad3-9a46-96179d59671e" containerID="9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.888706 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerDied","Data":"9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.900369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4" (OuterVolumeSpecName: "kube-api-access-8gkl4") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "kube-api-access-8gkl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.900743 4907 generic.go:334] "Generic (PLEG): container finished" podID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerID="c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d" exitCode=143 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.900916 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerDied","Data":"c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.923126 4907 generic.go:334] "Generic (PLEG): container finished" podID="516d1499-3461-4af5-b426-9ae4a711a468" containerID="73f68d1aced419107233520623fb182eb32be93775f62b9a147d1f6ec4671d78" exitCode=1 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.923187 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerDied","Data":"73f68d1aced419107233520623fb182eb32be93775f62b9a147d1f6ec4671d78"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.923216 4907 scope.go:117] "RemoveContainer" containerID="783ee4ed5f573bc2e551fccdb256e8146b5d15e716dc18f3c1e347aa5d0fa4c4" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.942741 4907 generic.go:334] "Generic (PLEG): container finished" podID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" exitCode=143 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.943000 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerDied","Data":"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.962664 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.963007 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.964738 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.964864 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.963138 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.965074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.965110 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.965123 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.980105 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.980446 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.980456 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8gkl4\" (UniqueName: \"kubernetes.io/projected/3793f53a-ca36-4b27-8444-d6dbfd860424-kube-api-access-8gkl4\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.984544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998272 4907 generic.go:334] "Generic (PLEG): container finished" podID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998637 4907 generic.go:334] "Generic (PLEG): container finished" podID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" exitCode=0 Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998594 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.998618 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerDied","Data":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.999801 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerDied","Data":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} Mar 13 14:30:43 crc kubenswrapper[4907]: I0313 14:30:43.999833 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dfdbcdd5-dcb7n" event={"ID":"1a1494c5-e2d6-4d29-b161-97b720742d8d","Type":"ContainerDied","Data":"3472978398de27f6988019a31485f47afeca847038681d998902a9798bd82c2a"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.016034 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data" (OuterVolumeSpecName: "config-data") pod "3793f53a-ca36-4b27-8444-d6dbfd860424" (UID: "3793f53a-ca36-4b27-8444-d6dbfd860424"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038242 4907 generic.go:334] "Generic (PLEG): container finished" podID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" exitCode=0 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038277 4907 generic.go:334] "Generic (PLEG): container finished" podID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" exitCode=0 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerDied","Data":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerDied","Data":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3793f53a-ca36-4b27-8444-d6dbfd860424","Type":"ContainerDied","Data":"8158569f0778590e4a7e4122294d9844f9402dad6a901826d7dd5af6e7fab659"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.038433 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.047074 4907 generic.go:334] "Generic (PLEG): container finished" podID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" exitCode=143 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.047180 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerDied","Data":"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.048706 4907 generic.go:334] "Generic (PLEG): container finished" podID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerID="f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927" exitCode=143 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.048768 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerDied","Data":"f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.049902 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-wvjr7" event={"ID":"82dec019-ba60-43f5-845b-322a6bcca225","Type":"ContainerStarted","Data":"cbbe85502bda2ac6ef5bc690e92165a5b92af3bcd3162f655a1bd43c95395317"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.056243 4907 generic.go:334] "Generic (PLEG): container finished" podID="001078cc-c01e-4173-9740-4f9ff082c593" containerID="3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0" exitCode=0 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.056337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerDied","Data":"3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.059772 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/ovsdbserver-sb/0.log" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.059875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"c2938e39-f1d8-4cdc-a32c-5d57b8f2034f","Type":"ContainerDied","Data":"c836a111eed36848627fde431a8ac117cdf68714d43b9b328bad3f2315e96e0e"} Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.059990 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.067406 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.072123 4907 generic.go:334] "Generic (PLEG): container finished" podID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" exitCode=143 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.072278 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerDied","Data":"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865"} Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.072628 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.079706 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.079761 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.081817 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.081917 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3793f53a-ca36-4b27-8444-d6dbfd860424-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.103406 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.109099 4907 scope.go:117] "RemoveContainer" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.145333 4907 scope.go:117] "RemoveContainer" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.181700 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.187957 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-7dfdbcdd5-dcb7n"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.202933 4907 scope.go:117] "RemoveContainer" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.204807 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.208347 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": container with ID starting with c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18 not found: ID does not exist" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.208380 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} err="failed to get container status \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": rpc error: code = NotFound desc = could not find container \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": container with ID starting with c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.208413 4907 scope.go:117] "RemoveContainer" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.209073 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": container with ID starting with 676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4 not found: ID does not exist" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209358 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} err="failed to get container status \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": rpc error: code = NotFound desc = could not find container \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": container with ID starting with 676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209379 4907 scope.go:117] "RemoveContainer" containerID="c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209585 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18"} err="failed to get container status \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": rpc error: code = NotFound desc = could not find container \"c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18\": container with ID starting with c6c4a6a54081de702756857f0341351ea26c5f9aaa725027df2c989c6b1a6e18 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.209603 4907 scope.go:117] "RemoveContainer" containerID="676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.212349 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4"} err="failed to get container status \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": rpc error: code = NotFound desc = could not find container \"676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4\": container with ID starting with 676efd80769fe771754cb647e65098f243a686b0c64467414fa0d6ccfbece2d4 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.212394 4907 scope.go:117] "RemoveContainer" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.213124 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.221908 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.230022 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.231307 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.232506 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.233473 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.233500 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.234676 4907 scope.go:117] "RemoveContainer" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.262182 4907 scope.go:117] "RemoveContainer" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.272630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": container with ID starting with 94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749 not found: ID does not exist" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.272670 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} err="failed to get container status \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": rpc error: code = NotFound desc = could not find container \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": container with ID starting with 94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.272700 4907 scope.go:117] "RemoveContainer" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.280145 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": container with ID starting with d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb not found: ID does not exist" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.280192 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} err="failed to get container status \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": rpc error: code = NotFound desc = could not find container \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": container with ID starting with d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.280219 4907 scope.go:117] "RemoveContainer" containerID="94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.282761 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749"} err="failed to get container status \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": rpc error: code = NotFound desc = could not find container \"94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749\": container with ID starting with 94bc55f3afe5f98100548e7593f16440afc25ced9f4af83fb6cd9f896253e749 not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.282807 4907 scope.go:117] "RemoveContainer" containerID="d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.285389 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.285672 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" containerID="cri-o://285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.286135 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" containerID="cri-o://16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.286190 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" containerID="cri-o://78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.286188 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" containerID="cri-o://79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.287349 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb"} err="failed to get container status \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": rpc error: code = NotFound desc = could not find container \"d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb\": container with ID starting with d484e26e1510dc0830d22229bd3bbd9bfa0e7d5cab421e8e5b87043a8df1abeb not found: ID does not exist" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.287403 4907 scope.go:117] "RemoveContainer" containerID="1ff45617937d399a62e2b1c7ef156ccf7052d51f57fba51d267daa3e49810ce6" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295272 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295354 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295382 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.295488 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") pod \"62983e04-f043-4ad3-9a46-96179d59671e\" (UID: \"62983e04-f043-4ad3-9a46-96179d59671e\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.319616 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq" (OuterVolumeSpecName: "kube-api-access-6kqzq") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "kube-api-access-6kqzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.339463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.383822 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.384131 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" containerID="cri-o://0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.410113 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kqzq\" (UniqueName: \"kubernetes.io/projected/62983e04-f043-4ad3-9a46-96179d59671e-kube-api-access-6kqzq\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.410168 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410271 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410333 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:48.410318135 +0000 UTC m=+1547.310105824 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410778 4907 configmap.go:193] Couldn't get configMap openstack/openstack-cell1-scripts: configmap "openstack-cell1-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.410818 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts podName:516d1499-3461-4af5-b426-9ae4a711a468 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:46.410805297 +0000 UTC m=+1545.310592986 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts") pod "root-account-create-update-glzbl" (UID: "516d1499-3461-4af5-b426-9ae4a711a468") : configmap "openstack-cell1-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.426754 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.440868 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.441579 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.463211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data" (OuterVolumeSpecName: "config-data") pod "62983e04-f043-4ad3-9a46-96179d59671e" (UID: "62983e04-f043-4ad3-9a46-96179d59671e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.468553 4907 scope.go:117] "RemoveContainer" containerID="432aefb6d80f0522ac7d6d516626fd7ea34147a98d43c3ddba905f2a1bb074ad" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.475053 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.475299 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" containerID="cri-o://65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.495962 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.511737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") pod \"bc15f080-f90d-4e97-a866-d5edfd662bb7\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.511900 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") pod \"bc15f080-f90d-4e97-a866-d5edfd662bb7\" (UID: \"bc15f080-f90d-4e97-a866-d5edfd662bb7\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.512412 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.512427 4907 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.512438 4907 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/62983e04-f043-4ad3-9a46-96179d59671e-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.517056 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc15f080-f90d-4e97-a866-d5edfd662bb7" (UID: "bc15f080-f90d-4e97-a866-d5edfd662bb7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.519374 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-04a8-account-create-update-ws5ht"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5" (OuterVolumeSpecName: "kube-api-access-jstl5") pod "bc15f080-f90d-4e97-a866-d5edfd662bb7" (UID: "bc15f080-f90d-4e97-a866-d5edfd662bb7"). InnerVolumeSpecName "kube-api-access-jstl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528514 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528870 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528898 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528911 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528917 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528935 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528941 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528953 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528959 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528968 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528974 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528981 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.528987 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.528999 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529005 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529012 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529017 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529027 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529033 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529043 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529049 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="init" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529061 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="init" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529068 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529075 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529086 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529092 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.529102 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529109 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529270 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37978cc1-fcba-4032-a8b1-6632b61692ff" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529282 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="ovsdbserver-nb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529291 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da13722-732a-4b65-a894-fe4612f30d75" containerName="dnsmasq-dns" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529300 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-server" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529308 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529318 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529326 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" containerName="ovsdbserver-sb" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529336 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" containerName="proxy-httpd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529345 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="62983e04-f043-4ad3-9a46-96179d59671e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529353 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="056dd756-0d7b-471b-9929-f622d05ad606" containerName="openstack-network-exporter" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529362 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="probe" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529375 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" containerName="cinder-scheduler" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.529383 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="26827945-75f0-4867-ba04-31ff6428e06a" containerName="ovn-northd" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.530046 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.535558 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.549685 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.580593 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.597970 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.613975 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-kcjn8"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.615481 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.615695 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.616367 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jstl5\" (UniqueName: \"kubernetes.io/projected/bc15f080-f90d-4e97-a866-d5edfd662bb7-kube-api-access-jstl5\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.616494 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc15f080-f90d-4e97-a866-d5edfd662bb7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.628647 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-crrww"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.665320 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.665550 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-85b9d64669-dxnf6" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" containerID="cri-o://e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.679206 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.684533 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.711233 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.720677 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") pod \"82dec019-ba60-43f5-845b-322a6bcca225\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.720840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") pod \"82dec019-ba60-43f5-845b-322a6bcca225\" (UID: \"82dec019-ba60-43f5-845b-322a6bcca225\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.721303 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.721348 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.721553 4907 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.721605 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:45.221588078 +0000 UTC m=+1544.121375767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : configmap "openstack-scripts" not found Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.722278 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "82dec019-ba60-43f5-845b-322a6bcca225" (UID: "82dec019-ba60-43f5-845b-322a6bcca225"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.726536 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv" (OuterVolumeSpecName: "kube-api-access-vlhjv") pod "82dec019-ba60-43f5-845b-322a6bcca225" (UID: "82dec019-ba60-43f5-845b-322a6bcca225"). InnerVolumeSpecName "kube-api-access-vlhjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.726731 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.733136 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.733577 4907 projected.go:194] Error preparing data for projected volume kube-api-access-48ngf for pod openstack/keystone-04a8-account-create-update-kntb5: failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.733639 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:45.233621365 +0000 UTC m=+1544.133409064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-48ngf" (UniqueName: "kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:44 crc kubenswrapper[4907]: E0313 14:30:44.733910 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-48ngf operator-scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone-04a8-account-create-update-kntb5" podUID="4685651c-2690-48f8-97d2-8817fedaff56" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.749181 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-hg688"] Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.750640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822498 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822590 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822730 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822841 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") pod \"516d1499-3461-4af5-b426-9ae4a711a468\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822914 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") pod \"516d1499-3461-4af5-b426-9ae4a711a468\" (UID: \"516d1499-3461-4af5-b426-9ae4a711a468\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822948 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.822975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823003 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823042 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") pod \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\" (UID: \"58ab9689-3ab2-42f2-ad56-beb22e29f8da\") " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823822 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82dec019-ba60-43f5-845b-322a6bcca225-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.823842 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlhjv\" (UniqueName: \"kubernetes.io/projected/82dec019-ba60-43f5-845b-322a6bcca225-kube-api-access-vlhjv\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.827211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.827673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.827687 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.828347 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "516d1499-3461-4af5-b426-9ae4a711a468" (UID: "516d1499-3461-4af5-b426-9ae4a711a468"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.828479 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.834396 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4" (OuterVolumeSpecName: "kube-api-access-kvwz4") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "kube-api-access-kvwz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.834623 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597" (OuterVolumeSpecName: "kube-api-access-df597") pod "516d1499-3461-4af5-b426-9ae4a711a468" (UID: "516d1499-3461-4af5-b426-9ae4a711a468"). InnerVolumeSpecName "kube-api-access-df597". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.866751 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "mysql-db") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.877263 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.919561 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "58ab9689-3ab2-42f2-ad56-beb22e29f8da" (UID: "58ab9689-3ab2-42f2-ad56-beb22e29f8da"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.932465 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" containerID="cri-o://b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa" gracePeriod=30 Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.933774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934616 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934641 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvwz4\" (UniqueName: \"kubernetes.io/projected/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kube-api-access-kvwz4\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934651 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-default\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934661 4907 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/58ab9689-3ab2-42f2-ad56-beb22e29f8da-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934669 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/516d1499-3461-4af5-b426-9ae4a711a468-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934677 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df597\" (UniqueName: \"kubernetes.io/projected/516d1499-3461-4af5-b426-9ae4a711a468-kube-api-access-df597\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934699 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934708 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934731 4907 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/58ab9689-3ab2-42f2-ad56-beb22e29f8da-kolla-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.934742 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/58ab9689-3ab2-42f2-ad56-beb22e29f8da-config-data-generated\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.947132 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:44 crc kubenswrapper[4907]: I0313 14:30:44.969395 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.032145 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" probeResult="failure" output="" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.035967 4907 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Mar 13 14:30:45 crc kubenswrapper[4907]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2026-03-13T14:30:42Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Mar 13 14:30:45 crc kubenswrapper[4907]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Mar 13 14:30:45 crc kubenswrapper[4907]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-bw2xb" message=< Mar 13 14:30:45 crc kubenswrapper[4907]: Exiting ovn-controller (1) [FAILED] Mar 13 14:30:45 crc kubenswrapper[4907]: Killing ovn-controller (1) [ OK ] Mar 13 14:30:45 crc kubenswrapper[4907]: 2026-03-13T14:30:42Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Mar 13 14:30:45 crc kubenswrapper[4907]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Mar 13 14:30:45 crc kubenswrapper[4907]: > Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.036010 4907 kuberuntime_container.go:691] "PreStop hook failed" err=< Mar 13 14:30:45 crc kubenswrapper[4907]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2026-03-13T14:30:42Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Mar 13 14:30:45 crc kubenswrapper[4907]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Mar 13 14:30:45 crc kubenswrapper[4907]: > pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" containerID="cri-o://df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036043 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-bw2xb" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" containerID="cri-o://df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" gracePeriod=27 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") pod \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036806 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") pod \"e8f6894c-b714-431e-912d-8b8de1a3914b\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036875 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") pod \"e8f6894c-b714-431e-912d-8b8de1a3914b\" (UID: \"e8f6894c-b714-431e-912d-8b8de1a3914b\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.036941 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") pod \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\" (UID: \"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.037233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" (UID: "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.037417 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.037432 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.038332 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8f6894c-b714-431e-912d-8b8de1a3914b" (UID: "e8f6894c-b714-431e-912d-8b8de1a3914b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.041498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65" (OuterVolumeSpecName: "kube-api-access-kfp65") pod "e8f6894c-b714-431e-912d-8b8de1a3914b" (UID: "e8f6894c-b714-431e-912d-8b8de1a3914b"). InnerVolumeSpecName "kube-api-access-kfp65". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.047066 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx" (OuterVolumeSpecName: "kube-api-access-wqnkx") pod "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" (UID: "4a4c64a3-7378-4d8d-9cf3-17583c6a90c0"). InnerVolumeSpecName "kube-api-access-wqnkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.098800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"62983e04-f043-4ad3-9a46-96179d59671e","Type":"ContainerDied","Data":"1cd31d9d43a5e6a87bb0836d84499cc5b9fab420b814ef48408f1a9549204fd0"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.098854 4907 scope.go:117] "RemoveContainer" containerID="9e2de4df42a807d1995b769c3c69187ddaaf7da1c1d09354b07088fb415f36d7" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.098993 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.107056 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-17a8-account-create-update-nj9pw" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.107711 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-17a8-account-create-update-nj9pw" event={"ID":"bc15f080-f90d-4e97-a866-d5edfd662bb7","Type":"ContainerDied","Data":"bc3dc3b0fd90acb1ff4424b36cbf0bd4f5dd344f9b23a791c924f552a3241a63"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.125857 4907 generic.go:334] "Generic (PLEG): container finished" podID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" exitCode=0 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.125934 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerDied","Data":"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.125961 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"58ab9689-3ab2-42f2-ad56-beb22e29f8da","Type":"ContainerDied","Data":"be11df2bf36396e7ed2d074eed97b3e79df7e43e0e8a1a8dec1ef8739586a724"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.126019 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.138519 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfp65\" (UniqueName: \"kubernetes.io/projected/e8f6894c-b714-431e-912d-8b8de1a3914b-kube-api-access-kfp65\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.138547 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8f6894c-b714-431e-912d-8b8de1a3914b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.138556 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqnkx\" (UniqueName: \"kubernetes.io/projected/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0-kube-api-access-wqnkx\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.139561 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-glzbl" event={"ID":"516d1499-3461-4af5-b426-9ae4a711a468","Type":"ContainerDied","Data":"f6d2f722265ae72bf48903452fd81587c07136e043b3e098f648b2d79cb3fd59"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.139631 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-glzbl" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.147154 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149424 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" exitCode=0 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149446 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" exitCode=2 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149454 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" exitCode=0 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149802 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.149815 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.152815 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9c64-account-create-update-5vvgg" event={"ID":"4a4c64a3-7378-4d8d-9cf3-17583c6a90c0","Type":"ContainerDied","Data":"a22452114ba800cf0f41433cad0c073716fcc8d240adefbf18595bc1cdb4a098"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.152865 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9c64-account-create-update-5vvgg" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158153 4907 generic.go:334] "Generic (PLEG): container finished" podID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" exitCode=2 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158195 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerDied","Data":"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158212 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c49286b1-2e9f-4de6-b7de-2e952f3ad607","Type":"ContainerDied","Data":"50ab44d5af6ebe42498682725ea848a78fc975fa9daf535d39dda1ec402423d1"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.158265 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.159931 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b69e-account-create-update-96wq7" event={"ID":"e8f6894c-b714-431e-912d-8b8de1a3914b","Type":"ContainerDied","Data":"d38b680247762c5b2510a8776b9df1c80f0da04bd9de35e5d2db2452cf7446f0"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.160037 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b69e-account-create-update-96wq7" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.161357 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bw2xb_6f8119b2-e38a-494d-967f-5198b83512c7/ovn-controller/0.log" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.161386 4907 generic.go:334] "Generic (PLEG): container finished" podID="6f8119b2-e38a-494d-967f-5198b83512c7" containerID="df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" exitCode=143 Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.161424 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerDied","Data":"df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.171632 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-47e2-account-create-update-wvjr7" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.172061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-47e2-account-create-update-wvjr7" event={"ID":"82dec019-ba60-43f5-845b-322a6bcca225","Type":"ContainerDied","Data":"cbbe85502bda2ac6ef5bc690e92165a5b92af3bcd3162f655a1bd43c95395317"} Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.172568 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.187315 4907 scope.go:117] "RemoveContainer" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.199732 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.201159 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.224923 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239709 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239761 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239833 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.239967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") pod \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\" (UID: \"c49286b1-2e9f-4de6-b7de-2e952f3ad607\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.240216 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.240405 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.240407 4907 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.240493 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:46.240470876 +0000 UTC m=+1545.140258635 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : configmap "openstack-scripts" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.250783 4907 projected.go:194] Error preparing data for projected volume kube-api-access-48ngf for pod openstack/keystone-04a8-account-create-update-kntb5: failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.250865 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:46.250843999 +0000 UTC m=+1545.150631698 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-48ngf" (UniqueName: "kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.259585 4907 scope.go:117] "RemoveContainer" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.278111 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.280899 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv" (OuterVolumeSpecName: "kube-api-access-29ctv") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "kube-api-access-29ctv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.286621 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-17a8-account-create-update-nj9pw"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.307017 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.342149 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29ctv\" (UniqueName: \"kubernetes.io/projected/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-api-access-29ctv\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.348252 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.356788 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.380251 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.384230 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.392142 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c49286b1-2e9f-4de6-b7de-2e952f3ad607" (UID: "c49286b1-2e9f-4de6-b7de-2e952f3ad607"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.392215 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-47e2-account-create-update-wvjr7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.406002 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bw2xb_6f8119b2-e38a-494d-967f-5198b83512c7/ovn-controller/0.log" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.406086 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.413330 4907 scope.go:117] "RemoveContainer" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.414314 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a\": container with ID starting with af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a not found: ID does not exist" containerID="af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.414359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a"} err="failed to get container status \"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a\": rpc error: code = NotFound desc = could not find container \"af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a\": container with ID starting with af7b1c6225d2e94f5a2f9827f7b02dc18651ab20c1d53b4a73e6d032e463897a not found: ID does not exist" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.414385 4907 scope.go:117] "RemoveContainer" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.415231 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86\": container with ID starting with 62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86 not found: ID does not exist" containerID="62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.415350 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86"} err="failed to get container status \"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86\": rpc error: code = NotFound desc = could not find container \"62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86\": container with ID starting with 62c42ba7e98c9bfe072975553c668c323179c088fa7a92a9812e075126aebb86 not found: ID does not exist" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.415471 4907 scope.go:117] "RemoveContainer" containerID="73f68d1aced419107233520623fb182eb32be93775f62b9a147d1f6ec4671d78" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.421149 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.439464 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-glzbl"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444128 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444197 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444264 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444293 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444386 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run" (OuterVolumeSpecName: "var-run") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444475 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.444543 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") pod \"6f8119b2-e38a-494d-967f-5198b83512c7\" (UID: \"6f8119b2-e38a-494d-967f-5198b83512c7\") " Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445304 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445323 4907 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445334 4907 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445343 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c49286b1-2e9f-4de6-b7de-2e952f3ad607-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445532 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.445927 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.447602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts" (OuterVolumeSpecName: "scripts") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.454475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5" (OuterVolumeSpecName: "kube-api-access-v2qq5") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "kube-api-access-v2qq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.457010 4907 scope.go:117] "RemoveContainer" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.481826 4907 scope.go:117] "RemoveContainer" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.483125 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.483211 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589\": container with ID starting with 0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589 not found: ID does not exist" containerID="0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.483279 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589"} err="failed to get container status \"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589\": rpc error: code = NotFound desc = could not find container \"0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589\": container with ID starting with 0d242394ff0714a798d1ef385dde5709185206b531dfb7b51abcf0db5d059589 not found: ID does not exist" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.508423 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.518180 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9c64-account-create-update-5vvgg"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.528757 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "6f8119b2-e38a-494d-967f-5198b83512c7" (UID: "6f8119b2-e38a-494d-967f-5198b83512c7"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.534263 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.543451 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b69e-account-create-update-96wq7"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547509 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547539 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8119b2-e38a-494d-967f-5198b83512c7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547550 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547560 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8119b2-e38a-494d-967f-5198b83512c7-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547569 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2qq5\" (UniqueName: \"kubernetes.io/projected/6f8119b2-e38a-494d-967f-5198b83512c7-kube-api-access-v2qq5\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.547577 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f8119b2-e38a-494d-967f-5198b83512c7-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.555480 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.560449 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.599503 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.172:8776/healthcheck\": read tcp 10.217.0.2:39704->10.217.0.172:8776: read: connection reset by peer" Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.752226 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:45 crc kubenswrapper[4907]: E0313 14:30:45.752296 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:49.752279757 +0000 UTC m=+1548.652067436 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.799830 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a1494c5-e2d6-4d29-b161-97b720742d8d" path="/var/lib/kubelet/pods/1a1494c5-e2d6-4d29-b161-97b720742d8d/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.800854 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.208:3000/\": dial tcp 10.217.0.208:3000: connect: connection refused" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.806518 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a0ea40-bdaa-4eef-8a28-263d8db74f10" path="/var/lib/kubelet/pods/34a0ea40-bdaa-4eef-8a28-263d8db74f10/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.807551 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3793f53a-ca36-4b27-8444-d6dbfd860424" path="/var/lib/kubelet/pods/3793f53a-ca36-4b27-8444-d6dbfd860424/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.824556 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ba2d41-5de4-48cb-aad6-59df13307ed1" path="/var/lib/kubelet/pods/37ba2d41-5de4-48cb-aad6-59df13307ed1/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.826520 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480d23fe-2980-46b3-9e86-e74f2d051357" path="/var/lib/kubelet/pods/480d23fe-2980-46b3-9e86-e74f2d051357/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.827207 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a4c64a3-7378-4d8d-9cf3-17583c6a90c0" path="/var/lib/kubelet/pods/4a4c64a3-7378-4d8d-9cf3-17583c6a90c0/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.827790 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="516d1499-3461-4af5-b426-9ae4a711a468" path="/var/lib/kubelet/pods/516d1499-3461-4af5-b426-9ae4a711a468/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.836529 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" path="/var/lib/kubelet/pods/58ab9689-3ab2-42f2-ad56-beb22e29f8da/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.838766 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62983e04-f043-4ad3-9a46-96179d59671e" path="/var/lib/kubelet/pods/62983e04-f043-4ad3-9a46-96179d59671e/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.840932 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82dec019-ba60-43f5-845b-322a6bcca225" path="/var/lib/kubelet/pods/82dec019-ba60-43f5-845b-322a6bcca225/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.841542 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c985ff9-3545-4544-acc8-ca9eb2bfad6a" path="/var/lib/kubelet/pods/9c985ff9-3545-4544-acc8-ca9eb2bfad6a/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.846517 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a055ce7f-0538-4e5e-938a-eaf6d0aa2dec" path="/var/lib/kubelet/pods/a055ce7f-0538-4e5e-938a-eaf6d0aa2dec/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.848832 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc15f080-f90d-4e97-a866-d5edfd662bb7" path="/var/lib/kubelet/pods/bc15f080-f90d-4e97-a866-d5edfd662bb7/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.852983 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2938e39-f1d8-4cdc-a32c-5d57b8f2034f" path="/var/lib/kubelet/pods/c2938e39-f1d8-4cdc-a32c-5d57b8f2034f/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.854522 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" path="/var/lib/kubelet/pods/c49286b1-2e9f-4de6-b7de-2e952f3ad607/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.867724 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8f6894c-b714-431e-912d-8b8de1a3914b" path="/var/lib/kubelet/pods/e8f6894c-b714-431e-912d-8b8de1a3914b/volumes" Mar 13 14:30:45 crc kubenswrapper[4907]: I0313 14:30:45.946435 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058526 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058586 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058606 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058649 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.058695 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") pod \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\" (UID: \"e3ee069a-41fb-4cab-b650-9bb9c03ac271\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.059698 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs" (OuterVolumeSpecName: "logs") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.067437 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t" (OuterVolumeSpecName: "kube-api-access-gds2t") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "kube-api-access-gds2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.069648 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts" (OuterVolumeSpecName: "scripts") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.126919 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data" (OuterVolumeSpecName: "config-data") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160447 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160464 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160474 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gds2t\" (UniqueName: \"kubernetes.io/projected/e3ee069a-41fb-4cab-b650-9bb9c03ac271-kube-api-access-gds2t\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160484 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.160495 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3ee069a-41fb-4cab-b650-9bb9c03ac271-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.164415 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.181227 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e3ee069a-41fb-4cab-b650-9bb9c03ac271" (UID: "e3ee069a-41fb-4cab-b650-9bb9c03ac271"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.184454 4907 generic.go:334] "Generic (PLEG): container finished" podID="4251ae36-90ad-41ea-915e-862df60f5c07" containerID="479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.184551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerDied","Data":"479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191489 4907 generic.go:334] "Generic (PLEG): container finished" podID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerID="65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerDied","Data":"65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191620 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b","Type":"ContainerDied","Data":"c174ea292d3346503480cf55d66ad04b5cc26c89fac549263ee565359ea5a081"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.191632 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c174ea292d3346503480cf55d66ad04b5cc26c89fac549263ee565359ea5a081" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.194385 4907 generic.go:334] "Generic (PLEG): container finished" podID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerID="ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.194449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerDied","Data":"ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196217 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-bw2xb_6f8119b2-e38a-494d-967f-5198b83512c7/ovn-controller/0.log" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196346 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-bw2xb" event={"ID":"6f8119b2-e38a-494d-967f-5198b83512c7","Type":"ContainerDied","Data":"4f06fa5373719502d5e6526674b7a5bdf213bd13582c27f9af62c582539db3b6"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196436 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-bw2xb" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.196445 4907 scope.go:117] "RemoveContainer" containerID="df7f17b832ba26f166592b0696a663ded9a2e03678601f86699f09b76e050e58" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.201099 4907 generic.go:334] "Generic (PLEG): container finished" podID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerID="0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.201174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerDied","Data":"0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203595 4907 generic.go:334] "Generic (PLEG): container finished" podID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerID="a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203676 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerDied","Data":"a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203730 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76","Type":"ContainerDied","Data":"10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.203742 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10db10f108be05500948726431cb9db777e566b5230954b5ba74892f495e9236" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206044 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerDied","Data":"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-85dc486cc4-fwrm6" event={"ID":"e3ee069a-41fb-4cab-b650-9bb9c03ac271","Type":"ContainerDied","Data":"80317f4f67489584d7a194b5c3eced56f03150f3c1e4e1db7ac581e0ad05f710"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.206141 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-85dc486cc4-fwrm6" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.210520 4907 generic.go:334] "Generic (PLEG): container finished" podID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerID="22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07" exitCode=0 Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.210614 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.210925 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerDied","Data":"22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07"} Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.262858 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.262936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") pod \"keystone-04a8-account-create-update-kntb5\" (UID: \"4685651c-2690-48f8-97d2-8817fedaff56\") " pod="openstack/keystone-04a8-account-create-update-kntb5" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.263164 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.263187 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e3ee069a-41fb-4cab-b650-9bb9c03ac271-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.263251 4907 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.263316 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:48.263300808 +0000 UTC m=+1547.163088497 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : configmap "openstack-scripts" not found Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.268574 4907 projected.go:194] Error preparing data for projected volume kube-api-access-48ngf for pod openstack/keystone-04a8-account-create-update-kntb5: failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.268690 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf podName:4685651c-2690-48f8-97d2-8817fedaff56 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:48.26867153 +0000 UTC m=+1547.168459219 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-48ngf" (UniqueName: "kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf") pod "keystone-04a8-account-create-update-kntb5" (UID: "4685651c-2690-48f8-97d2-8817fedaff56") : failed to fetch token: serviceaccounts "galera-openstack" not found Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.345952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.351577 4907 scope.go:117] "RemoveContainer" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.376036 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.398967 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.400587 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.405361 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-04a8-account-create-update-kntb5"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.406025 4907 scope.go:117] "RemoveContainer" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.427722 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.436224 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.439213 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.444105 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-bw2xb"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.447187 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.456971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.459838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-85dc486cc4-fwrm6"] Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470522 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470581 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470620 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470650 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470678 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470733 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470807 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470829 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470927 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.470996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471021 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471044 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471072 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471096 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471152 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") pod \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\" (UID: \"cacb281b-ee3b-42b6-8754-d07e6ea6aa4b\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471178 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471210 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471267 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471302 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471360 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471461 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471489 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471570 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471618 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") pod \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\" (UID: \"2e41272f-6e7b-47a7-825e-3571f6a1fd07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471642 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"4251ae36-90ad-41ea-915e-862df60f5c07\" (UID: \"4251ae36-90ad-41ea-915e-862df60f5c07\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.471663 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") pod \"0b2e62fa-af35-4f35-8852-362ea97c17e7\" (UID: \"0b2e62fa-af35-4f35-8852-362ea97c17e7\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.472048 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48ngf\" (UniqueName: \"kubernetes.io/projected/4685651c-2690-48f8-97d2-8817fedaff56-kube-api-access-48ngf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.472076 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4685651c-2690-48f8-97d2-8817fedaff56-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.473016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.476140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs" (OuterVolumeSpecName: "logs") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.483504 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk" (OuterVolumeSpecName: "kube-api-access-8rvmk") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "kube-api-access-8rvmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.488848 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.493930 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.499347 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.499466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f" (OuterVolumeSpecName: "kube-api-access-vdj7f") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "kube-api-access-vdj7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.501391 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs" (OuterVolumeSpecName: "logs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.505135 4907 scope.go:117] "RemoveContainer" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.507101 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts" (OuterVolumeSpecName: "scripts") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.508478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data" (OuterVolumeSpecName: "config-data") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.508519 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.510131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs" (OuterVolumeSpecName: "logs") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.510233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.511596 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs" (OuterVolumeSpecName: "logs") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.511721 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts" (OuterVolumeSpecName: "scripts") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.511922 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e\": container with ID starting with 6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e not found: ID does not exist" containerID="6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.511994 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e"} err="failed to get container status \"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e\": rpc error: code = NotFound desc = could not find container \"6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e\": container with ID starting with 6d08d480e1be3a68c3d20d884244fa154245d26dcb3fb269a2aeb6806364407e not found: ID does not exist" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.512033 4907 scope.go:117] "RemoveContainer" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.513218 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj" (OuterVolumeSpecName: "kube-api-access-2jswj") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "kube-api-access-2jswj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.514100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph" (OuterVolumeSpecName: "kube-api-access-h2bph") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "kube-api-access-h2bph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.514107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts" (OuterVolumeSpecName: "scripts") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.514182 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: E0313 14:30:46.516266 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89\": container with ID starting with c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89 not found: ID does not exist" containerID="c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.516302 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89"} err="failed to get container status \"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89\": rpc error: code = NotFound desc = could not find container \"c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89\": container with ID starting with c0c472fe526cce04d2912530f4de0e17fa4ad4be341dbd2a462b13899fc1ad89 not found: ID does not exist" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.518593 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6" (OuterVolumeSpecName: "kube-api-access-dnjx6") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "kube-api-access-dnjx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.554794 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.565004 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575442 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575484 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575532 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575631 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575719 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.575760 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") pod \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\" (UID: \"1b2f62d4-0584-4cc3-81ed-e067d8db23b1\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576140 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576157 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576166 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576176 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2bph\" (UniqueName: \"kubernetes.io/projected/4251ae36-90ad-41ea-915e-862df60f5c07-kube-api-access-h2bph\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576188 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e41272f-6e7b-47a7-825e-3571f6a1fd07-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576197 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576208 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rvmk\" (UniqueName: \"kubernetes.io/projected/2e41272f-6e7b-47a7-825e-3571f6a1fd07-kube-api-access-8rvmk\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576217 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576239 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576250 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576260 4907 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kolla-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576270 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdj7f\" (UniqueName: \"kubernetes.io/projected/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-kube-api-access-vdj7f\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576278 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576286 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576296 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4251ae36-90ad-41ea-915e-862df60f5c07-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576304 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jswj\" (UniqueName: \"kubernetes.io/projected/0b2e62fa-af35-4f35-8852-362ea97c17e7-kube-api-access-2jswj\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576312 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnjx6\" (UniqueName: \"kubernetes.io/projected/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-kube-api-access-dnjx6\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576321 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576329 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b2e62fa-af35-4f35-8852-362ea97c17e7-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576337 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576350 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.576358 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.581999 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs" (OuterVolumeSpecName: "logs") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.587694 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c" (OuterVolumeSpecName: "kube-api-access-ktf7c") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "kube-api-access-ktf7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.605307 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.614702 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.639397 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.640970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.651105 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.678666 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.678970 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687008 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687201 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktf7c\" (UniqueName: \"kubernetes.io/projected/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-kube-api-access-ktf7c\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687299 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687365 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.687422 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.705124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data" (OuterVolumeSpecName: "config-data") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.710317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.714418 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.725891 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.754776 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.758338 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.759129 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data" (OuterVolumeSpecName: "config-data") pod "0b2e62fa-af35-4f35-8852-362ea97c17e7" (UID: "0b2e62fa-af35-4f35-8852-362ea97c17e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.768960 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.773487 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2e41272f-6e7b-47a7-825e-3571f6a1fd07" (UID: "2e41272f-6e7b-47a7-825e-3571f6a1fd07"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.777040 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data" (OuterVolumeSpecName: "config-data") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.788049 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4251ae36-90ad-41ea-915e-862df60f5c07" (UID: "4251ae36-90ad-41ea-915e-862df60f5c07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.788616 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.788655 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.790059 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791092 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791153 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791176 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") pod \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\" (UID: \"41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791208 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") pod \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\" (UID: \"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d\") " Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791970 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.791994 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792007 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0b2e62fa-af35-4f35-8852-362ea97c17e7-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792019 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792030 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792041 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792051 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4251ae36-90ad-41ea-915e-862df60f5c07-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792062 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792076 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792087 4907 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e41272f-6e7b-47a7-825e-3571f6a1fd07-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.792659 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data" (OuterVolumeSpecName: "config-data") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: W0313 14:30:46.794066 4907 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76/volumes/kubernetes.io~secret/public-tls-certs Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.794110 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" (UID: "41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.794299 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs" (OuterVolumeSpecName: "logs") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.796078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg" (OuterVolumeSpecName: "kube-api-access-565tg") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "kube-api-access-565tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.802608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data" (OuterVolumeSpecName: "config-data") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.805642 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1b2f62d4-0584-4cc3-81ed-e067d8db23b1" (UID: "1b2f62d4-0584-4cc3-81ed-e067d8db23b1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.813646 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" (UID: "cacb281b-ee3b-42b6-8754-d07e6ea6aa4b"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.816606 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data" (OuterVolumeSpecName: "config-data") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.818755 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.838714 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.877919 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" (UID: "cc227d33-3bf1-4b5f-8368-8e6b8a512f7d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893906 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893930 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893942 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893952 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893960 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893968 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-565tg\" (UniqueName: \"kubernetes.io/projected/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-kube-api-access-565tg\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893977 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893985 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.893993 4907 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.894001 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1b2f62d4-0584-4cc3-81ed-e067d8db23b1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:46 crc kubenswrapper[4907]: I0313 14:30:46.894009 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.165209 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233028 4907 generic.go:334] "Generic (PLEG): container finished" podID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" exitCode=0 Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233100 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerDied","Data":"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" event={"ID":"65531461-cea2-4b2c-a9b6-8fd9e9bffb27","Type":"ContainerDied","Data":"d83bd2c5278d976f4863dc597e16a40e27298e4624048379804b2a946c5145cc"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233146 4907 scope.go:117] "RemoveContainer" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.233261 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7d7bc9f7d-fks76" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255721 4907 generic.go:334] "Generic (PLEG): container finished" podID="c4411bd2-c555-433a-9015-f623948b1401" containerID="b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa" exitCode=0 Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerDied","Data":"b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255831 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"c4411bd2-c555-433a-9015-f623948b1401","Type":"ContainerDied","Data":"da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.255846 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da35a91fc58ce0787798d501f92eec9bbafae6711281a9caca5018e3a65c78c8" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.266265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e41272f-6e7b-47a7-825e-3571f6a1fd07","Type":"ContainerDied","Data":"68545429a61b864d7541170bf482dacf433239060ec793e8fb62a4bc2ee9a2f3"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.266376 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302732 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302798 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.302959 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") pod \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\" (UID: \"65531461-cea2-4b2c-a9b6-8fd9e9bffb27\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.303566 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs" (OuterVolumeSpecName: "logs") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.331245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb" (OuterVolumeSpecName: "kube-api-access-lz7hb") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "kube-api-access-lz7hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.331793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0b2e62fa-af35-4f35-8852-362ea97c17e7","Type":"ContainerDied","Data":"0ede32ea598e7a7a739c54b190ba597842e8cec1fea7546f52eefbb2b1fa0f83"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.331968 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.359939 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.360519 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364088 4907 generic.go:334] "Generic (PLEG): container finished" podID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" exitCode=0 Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerDied","Data":"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364196 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cc227d33-3bf1-4b5f-8368-8e6b8a512f7d","Type":"ContainerDied","Data":"ed58fd3bb4c7be310b9f67af6133d99b09325a11963ab01983393f4f0a20f895"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.364260 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.367188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.377966 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4251ae36-90ad-41ea-915e-862df60f5c07","Type":"ContainerDied","Data":"d3f2a65d466df300e932d28e3a4a5fb0f68b272bc80ae37edb5a080aa7144cae"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.378077 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383200 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7fc86597fb-mt9pk" event={"ID":"1b2f62d4-0584-4cc3-81ed-e067d8db23b1","Type":"ContainerDied","Data":"97dce4dda51605af3970178d1030468b7ba5970d94173d175f92a65d74507fff"} Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383392 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7fc86597fb-mt9pk" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.383895 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.385466 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.403871 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404031 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404044 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404147 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404203 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.404688 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.405329 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.411602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data" (OuterVolumeSpecName: "config-data") pod "65531461-cea2-4b2c-a9b6-8fd9e9bffb27" (UID: "65531461-cea2-4b2c-a9b6-8fd9e9bffb27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412294 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") pod \"c4411bd2-c555-433a-9015-f623948b1401\" (UID: \"c4411bd2-c555-433a-9015-f623948b1401\") " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.412912 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.413832 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414324 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414353 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414365 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz7hb\" (UniqueName: \"kubernetes.io/projected/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-kube-api-access-lz7hb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414380 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414390 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65531461-cea2-4b2c-a9b6-8fd9e9bffb27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414399 4907 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-kolla-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414408 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/c4411bd2-c555-433a-9015-f623948b1401-config-data-generated\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414417 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-config-data-default\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.414425 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4411bd2-c555-433a-9015-f623948b1401-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.421627 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz" (OuterVolumeSpecName: "kube-api-access-pcwkz") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "kube-api-access-pcwkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.421834 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.439255 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.439399 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445084 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445479 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445531 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.445941 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.447047 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.447124 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.451002 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.457724 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.470264 4907 scope.go:117] "RemoveContainer" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.486096 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.498001 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.517920 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcwkz\" (UniqueName: \"kubernetes.io/projected/c4411bd2-c555-433a-9015-f623948b1401-kube-api-access-pcwkz\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.517957 4907 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.517993 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.523279 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4411bd2-c555-433a-9015-f623948b1401" (UID: "c4411bd2-c555-433a-9015-f623948b1401"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.526585 4907 scope.go:117] "RemoveContainer" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.527016 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4\": container with ID starting with 7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4 not found: ID does not exist" containerID="7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527040 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4"} err="failed to get container status \"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4\": rpc error: code = NotFound desc = could not find container \"7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4\": container with ID starting with 7117af57f620e2d772f6e2ca6d92edfdc283b3eb640ab414d9483ba1758f36d4 not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527056 4907 scope.go:117] "RemoveContainer" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.527613 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865\": container with ID starting with 3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865 not found: ID does not exist" containerID="3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527633 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865"} err="failed to get container status \"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865\": rpc error: code = NotFound desc = could not find container \"3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865\": container with ID starting with 3ccd8eb5dd33f0cece989e6411df4852cb55e2cb0b138f16332e7a07ccf5c865 not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.527647 4907 scope.go:117] "RemoveContainer" containerID="ef8644527a89477971dc8c42d7fc7d2d08c1853a9324c6451f42196c1c50329e" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.533965 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.539212 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.547750 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.553300 4907 scope.go:117] "RemoveContainer" containerID="c5ecad10cc9377aeda50af0d8c9213440ce021149e77309245ab84eb57179e9d" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.565416 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.587835 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.593284 4907 scope.go:117] "RemoveContainer" containerID="0d9941f62564b14687332e4dbc1f81a8df8a858860923a1b915fe67a744f75f5" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.594003 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.605832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.608129 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7fc86597fb-mt9pk"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.613471 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.618777 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.619001 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.619026 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4411bd2-c555-433a-9015-f623948b1401-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.623267 4907 scope.go:117] "RemoveContainer" containerID="eeaee513015ed79d41df919baf6b1bec0513b97e239abe794d310237db6cde32" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.629985 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.636777 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-7d7bc9f7d-fks76"] Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.664197 4907 scope.go:117] "RemoveContainer" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.687089 4907 scope.go:117] "RemoveContainer" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.707192 4907 scope.go:117] "RemoveContainer" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.711832 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd\": container with ID starting with 58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd not found: ID does not exist" containerID="58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.711896 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd"} err="failed to get container status \"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd\": rpc error: code = NotFound desc = could not find container \"58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd\": container with ID starting with 58121d89217f995e7e932419445b98334a3ccb8b202e18c097026d64131a91fd not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.711929 4907 scope.go:117] "RemoveContainer" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.712336 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd\": container with ID starting with 6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd not found: ID does not exist" containerID="6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.712360 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd"} err="failed to get container status \"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd\": rpc error: code = NotFound desc = could not find container \"6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd\": container with ID starting with 6c1fdac6667e489d30e980acdd57d1cf0fa750bc21c6f46f11adb992a74fe2bd not found: ID does not exist" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.712376 4907 scope.go:117] "RemoveContainer" containerID="479ee0fb693df48bcf5ef06770a45535f793e4153c556058631f42347318a544" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.740861 4907 scope.go:117] "RemoveContainer" containerID="70b1012f350567e4a184430af6b8324fc4a32cf4f648d2eb406bd3d58d2b10d8" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.762870 4907 scope.go:117] "RemoveContainer" containerID="22aa297a2f364f3884c53d61c34c94086546e7771e899b38ffebaf99570e4f07" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.784507 4907 scope.go:117] "RemoveContainer" containerID="f21714a9f3b46d6b404fb8600bc2b71af7d623654fb77b4f61140d6330569927" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.793916 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" path="/var/lib/kubelet/pods/0b2e62fa-af35-4f35-8852-362ea97c17e7/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.794597 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" path="/var/lib/kubelet/pods/1b2f62d4-0584-4cc3-81ed-e067d8db23b1/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.795234 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" path="/var/lib/kubelet/pods/2e41272f-6e7b-47a7-825e-3571f6a1fd07/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.796725 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" path="/var/lib/kubelet/pods/41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.797822 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" path="/var/lib/kubelet/pods/4251ae36-90ad-41ea-915e-862df60f5c07/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.798741 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4685651c-2690-48f8-97d2-8817fedaff56" path="/var/lib/kubelet/pods/4685651c-2690-48f8-97d2-8817fedaff56/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.801278 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" path="/var/lib/kubelet/pods/65531461-cea2-4b2c-a9b6-8fd9e9bffb27/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.802097 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" path="/var/lib/kubelet/pods/6f8119b2-e38a-494d-967f-5198b83512c7/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.803012 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" path="/var/lib/kubelet/pods/cacb281b-ee3b-42b6-8754-d07e6ea6aa4b/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.804197 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" path="/var/lib/kubelet/pods/cc227d33-3bf1-4b5f-8368-8e6b8a512f7d/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: I0313 14:30:47.804960 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" path="/var/lib/kubelet/pods/e3ee069a-41fb-4cab-b650-9bb9c03ac271/volumes" Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828551 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828587 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828598 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828609 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:47 crc kubenswrapper[4907]: E0313 14:30:47.828665 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:55.828646281 +0000 UTC m=+1554.728433970 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.215239 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.276765 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337566 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337604 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337692 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337712 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337770 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337843 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.337896 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") pod \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\" (UID: \"8a738a66-54bc-4f5e-86d5-6e3004d8f265\") " Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.339655 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.343098 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.343319 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb" (OuterVolumeSpecName: "kube-api-access-pvwdb") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "kube-api-access-pvwdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.346245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.349281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts" (OuterVolumeSpecName: "scripts") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.351308 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.355417 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.355569 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.374378 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data" (OuterVolumeSpecName: "config-data") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.376020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.424317 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427156 4907 generic.go:334] "Generic (PLEG): container finished" podID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" exitCode=0 Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerDied","Data":"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72"} Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427254 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85b9d64669-dxnf6" event={"ID":"8a738a66-54bc-4f5e-86d5-6e3004d8f265","Type":"ContainerDied","Data":"50cd2791b5f7c237a1259945e1c40427e424019970480fa895cfa345fab258b4"} Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427273 4907 scope.go:117] "RemoveContainer" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.427394 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85b9d64669-dxnf6" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.428697 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8a738a66-54bc-4f5e-86d5-6e3004d8f265" (UID: "8a738a66-54bc-4f5e-86d5-6e3004d8f265"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.429632 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452293 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452325 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452338 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452350 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452361 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.452359 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452376 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.452434 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data podName:286c82e8-b74a-49d7-a355-ac074aace10e nodeName:}" failed. No retries permitted until 2026-03-13 14:30:56.45241674 +0000 UTC m=+1555.352204429 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data") pod "rabbitmq-cell1-server-0" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e") : configmap "rabbitmq-cell1-config-data" not found Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452459 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a738a66-54bc-4f5e-86d5-6e3004d8f265-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.452475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvwdb\" (UniqueName: \"kubernetes.io/projected/8a738a66-54bc-4f5e-86d5-6e3004d8f265-kube-api-access-pvwdb\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.456557 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.461344 4907 scope.go:117] "RemoveContainer" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" Mar 13 14:30:48 crc kubenswrapper[4907]: E0313 14:30:48.461726 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72\": container with ID starting with e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72 not found: ID does not exist" containerID="e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.461784 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72"} err="failed to get container status \"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72\": rpc error: code = NotFound desc = could not find container \"e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72\": container with ID starting with e6a1933783ac7223012561b093ce055e734ac2c3c802390f6f0bf14354694b72 not found: ID does not exist" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.463246 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.696802 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.761836 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:30:48 crc kubenswrapper[4907]: I0313 14:30:48.766666 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-85b9d64669-dxnf6"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.054318 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.055943 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.057781 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.057821 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.231491 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.236199 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.240650 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.240692 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.334732 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.443631 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3c34454-315e-4821-ab25-b0f331a0d521" containerID="c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4" exitCode=0 Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.443678 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerDied","Data":"c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4"} Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446044 4907 generic.go:334] "Generic (PLEG): container finished" podID="286c82e8-b74a-49d7-a355-ac074aace10e" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" exitCode=0 Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446070 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerDied","Data":"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43"} Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446097 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"286c82e8-b74a-49d7-a355-ac074aace10e","Type":"ContainerDied","Data":"84aa3724ad3336159381b40fcc1f3bf130caa7609c9affeddec48f79002eb26c"} Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446114 4907 scope.go:117] "RemoveContainer" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.446216 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.471236 4907 scope.go:117] "RemoveContainer" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475723 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475819 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475859 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475971 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.475996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476030 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476050 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.476096 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") pod \"286c82e8-b74a-49d7-a355-ac074aace10e\" (UID: \"286c82e8-b74a-49d7-a355-ac074aace10e\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.477227 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.477243 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.477232 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.481211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.481554 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info" (OuterVolumeSpecName: "pod-info") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.481603 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.494674 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.497495 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq" (OuterVolumeSpecName: "kube-api-access-w8snq") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "kube-api-access-w8snq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508467 4907 scope.go:117] "RemoveContainer" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508562 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data" (OuterVolumeSpecName: "config-data") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.508898 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43\": container with ID starting with d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43 not found: ID does not exist" containerID="d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508931 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43"} err="failed to get container status \"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43\": rpc error: code = NotFound desc = could not find container \"d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43\": container with ID starting with d23153977130306af0e21b41c03dbbc96a217b011704059af7d65c0a4153cf43 not found: ID does not exist" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.508955 4907 scope.go:117] "RemoveContainer" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.509395 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7\": container with ID starting with 275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7 not found: ID does not exist" containerID="275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.509469 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7"} err="failed to get container status \"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7\": rpc error: code = NotFound desc = could not find container \"275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7\": container with ID starting with 275b76780d99e94e2c958194ce29198632d04fd0fe0f0668ff5025ad387fc2f7 not found: ID does not exist" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.526429 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf" (OuterVolumeSpecName: "server-conf") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578058 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578107 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578116 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/286c82e8-b74a-49d7-a355-ac074aace10e-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578126 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/286c82e8-b74a-49d7-a355-ac074aace10e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578155 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578166 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578178 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8snq\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-kube-api-access-w8snq\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578187 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578203 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.578212 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/286c82e8-b74a-49d7-a355-ac074aace10e-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.594093 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.597099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "286c82e8-b74a-49d7-a355-ac074aace10e" (UID: "286c82e8-b74a-49d7-a355-ac074aace10e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.680417 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.680444 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/286c82e8-b74a-49d7-a355-ac074aace10e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.753046 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.782203 4907 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Mar 13 14:30:49 crc kubenswrapper[4907]: E0313 14:30:49.782284 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data podName:e3c34454-315e-4821-ab25-b0f331a0d521 nodeName:}" failed. No retries permitted until 2026-03-13 14:30:57.782266645 +0000 UTC m=+1556.682054334 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data") pod "rabbitmq-server-0" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521") : configmap "rabbitmq-config-data" not found Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.805841 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" path="/var/lib/kubelet/pods/8a738a66-54bc-4f5e-86d5-6e3004d8f265/volumes" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.806466 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4411bd2-c555-433a-9015-f623948b1401" path="/var/lib/kubelet/pods/c4411bd2-c555-433a-9015-f623948b1401/volumes" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.809587 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.814366 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883665 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883869 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.883986 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884083 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884231 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884263 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884279 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884419 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884444 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") pod \"e3c34454-315e-4821-ab25-b0f331a0d521\" (UID: \"e3c34454-315e-4821-ab25-b0f331a0d521\") " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.884764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.885266 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.885292 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.887177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.887393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.887408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.889449 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l" (OuterVolumeSpecName: "kube-api-access-mnz7l") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "kube-api-access-mnz7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.890168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.890630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info" (OuterVolumeSpecName: "pod-info") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.903545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data" (OuterVolumeSpecName: "config-data") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.924096 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf" (OuterVolumeSpecName: "server-conf") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.968481 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e3c34454-315e-4821-ab25-b0f331a0d521" (UID: "e3c34454-315e-4821-ab25-b0f331a0d521"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988774 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988829 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988839 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988849 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnz7l\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-kube-api-access-mnz7l\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988858 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988867 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e3c34454-315e-4821-ab25-b0f331a0d521-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988874 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e3c34454-315e-4821-ab25-b0f331a0d521-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988897 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e3c34454-315e-4821-ab25-b0f331a0d521-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:49 crc kubenswrapper[4907]: I0313 14:30:49.988905 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e3c34454-315e-4821-ab25-b0f331a0d521-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.008873 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.090602 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.468160 4907 generic.go:334] "Generic (PLEG): container finished" podID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" exitCode=0 Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.468225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerDied","Data":"77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4"} Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.472034 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e3c34454-315e-4821-ab25-b0f331a0d521","Type":"ContainerDied","Data":"0dddf7e485bfeda5a825f940bee1a34df372878429700a5d41310e0d0412399b"} Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.472084 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.472102 4907 scope.go:117] "RemoveContainer" containerID="c5d00dd71aba57d230e51645212e11968b2acd8a912b082abd6373002fb907b4" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.506431 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.518557 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.522449 4907 scope.go:117] "RemoveContainer" containerID="987e5ce0ff0b4c3691f9df5efc31a14e40c988230f0cf14b4b06e00cd7129b38" Mar 13 14:30:50 crc kubenswrapper[4907]: I0313 14:30:50.883329 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.014670 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") pod \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.014746 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") pod \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.014781 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") pod \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\" (UID: \"25c1c7e6-a149-4a8e-aead-3ebdc7e56851\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.027100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v" (OuterVolumeSpecName: "kube-api-access-gxd7v") pod "25c1c7e6-a149-4a8e-aead-3ebdc7e56851" (UID: "25c1c7e6-a149-4a8e-aead-3ebdc7e56851"). InnerVolumeSpecName "kube-api-access-gxd7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.048329 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25c1c7e6-a149-4a8e-aead-3ebdc7e56851" (UID: "25c1c7e6-a149-4a8e-aead-3ebdc7e56851"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.059329 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data" (OuterVolumeSpecName: "config-data") pod "25c1c7e6-a149-4a8e-aead-3ebdc7e56851" (UID: "25c1c7e6-a149-4a8e-aead-3ebdc7e56851"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.116360 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxd7v\" (UniqueName: \"kubernetes.io/projected/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-kube-api-access-gxd7v\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.116396 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.116406 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25c1c7e6-a149-4a8e-aead-3ebdc7e56851-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.145228 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.155700 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.162002 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.162076 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.167:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.162285 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7fc86597fb-mt9pk" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.167:9311/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.256513 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.275574 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/memcached-0" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" probeResult="failure" output="dial tcp 10.217.0.108:11211: i/o timeout" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319147 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") pod \"4050d00b-0256-45c5-9dc4-0ab46956405d\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") pod \"4050d00b-0256-45c5-9dc4-0ab46956405d\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319294 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319320 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319342 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") pod \"4050d00b-0256-45c5-9dc4-0ab46956405d\" (UID: \"4050d00b-0256-45c5-9dc4-0ab46956405d\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319419 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319449 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319518 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319532 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319553 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319590 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") pod \"aa81fa2a-4c27-4e02-9891-5b39348feca9\" (UID: \"aa81fa2a-4c27-4e02-9891-5b39348feca9\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.319613 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") pod \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\" (UID: \"25b5fafe-65c8-48ba-bc34-83442f8ace4c\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.320247 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs" (OuterVolumeSpecName: "logs") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.321106 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.321970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.323446 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf" (OuterVolumeSpecName: "kube-api-access-99bgf") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "kube-api-access-99bgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24" (OuterVolumeSpecName: "kube-api-access-dvn24") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "kube-api-access-dvn24". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326376 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts" (OuterVolumeSpecName: "scripts") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326458 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.326713 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29" (OuterVolumeSpecName: "kube-api-access-xdg29") pod "4050d00b-0256-45c5-9dc4-0ab46956405d" (UID: "4050d00b-0256-45c5-9dc4-0ab46956405d"). InnerVolumeSpecName "kube-api-access-xdg29". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.346541 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4050d00b-0256-45c5-9dc4-0ab46956405d" (UID: "4050d00b-0256-45c5-9dc4-0ab46956405d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.347166 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.357062 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.357967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data" (OuterVolumeSpecName: "config-data") pod "4050d00b-0256-45c5-9dc4-0ab46956405d" (UID: "4050d00b-0256-45c5-9dc4-0ab46956405d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.366467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.372498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data" (OuterVolumeSpecName: "config-data") pod "25b5fafe-65c8-48ba-bc34-83442f8ace4c" (UID: "25b5fafe-65c8-48ba-bc34-83442f8ace4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.379682 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.399465 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data" (OuterVolumeSpecName: "config-data") pod "aa81fa2a-4c27-4e02-9891-5b39348feca9" (UID: "aa81fa2a-4c27-4e02-9891-5b39348feca9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.420978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") pod \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421033 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") pod \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421074 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") pod \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\" (UID: \"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec\") " Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421351 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421367 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99bgf\" (UniqueName: \"kubernetes.io/projected/25b5fafe-65c8-48ba-bc34-83442f8ace4c-kube-api-access-99bgf\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421377 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421385 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421397 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421404 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421412 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421422 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdg29\" (UniqueName: \"kubernetes.io/projected/4050d00b-0256-45c5-9dc4-0ab46956405d-kube-api-access-xdg29\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421433 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421444 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aa81fa2a-4c27-4e02-9891-5b39348feca9-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421454 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421462 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4050d00b-0256-45c5-9dc4-0ab46956405d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421470 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aa81fa2a-4c27-4e02-9891-5b39348feca9-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421477 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvn24\" (UniqueName: \"kubernetes.io/projected/aa81fa2a-4c27-4e02-9891-5b39348feca9-kube-api-access-dvn24\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421485 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25b5fafe-65c8-48ba-bc34-83442f8ace4c-logs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.421492 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/25b5fafe-65c8-48ba-bc34-83442f8ace4c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.424072 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj" (OuterVolumeSpecName: "kube-api-access-cdxcj") pod "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" (UID: "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec"). InnerVolumeSpecName "kube-api-access-cdxcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.442518 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data" (OuterVolumeSpecName: "config-data") pod "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" (UID: "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.453732 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" (UID: "364a702e-e6c6-42bc-8cee-4ba19ea2f1ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488399 4907 generic.go:334] "Generic (PLEG): container finished" podID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488510 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488539 4907 scope.go:117] "RemoveContainer" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.488527 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aa81fa2a-4c27-4e02-9891-5b39348feca9","Type":"ContainerDied","Data":"b4ef9b58ba405d6fb537a91498e7353e4bf4deae5a0e5f61d8548bb330be20cb"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.491559 4907 generic.go:334] "Generic (PLEG): container finished" podID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.491631 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerDied","Data":"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.491651 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"364a702e-e6c6-42bc-8cee-4ba19ea2f1ec","Type":"ContainerDied","Data":"416b8c99953bb29e1da48539894292de008485b785183dda880949342468afcc"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.492002 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.493156 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.494353 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"25c1c7e6-a149-4a8e-aead-3ebdc7e56851","Type":"ContainerDied","Data":"b71e79c4c15864ccbc5c66db3afab587a50c68e66b3d7994a1fc84ed59fd938f"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515294 4907 generic.go:334] "Generic (PLEG): container finished" podID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerDied","Data":"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515365 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.515377 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76bbcdf7d9-j6hw5" event={"ID":"25b5fafe-65c8-48ba-bc34-83442f8ace4c","Type":"ContainerDied","Data":"57bab78a65adf0bc133213358e1f66cc1c97a4b0aefeb088fc087562e2ca2999"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517435 4907 generic.go:334] "Generic (PLEG): container finished" podID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" exitCode=0 Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517461 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerDied","Data":"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4050d00b-0256-45c5-9dc4-0ab46956405d","Type":"ContainerDied","Data":"c18abca4f12b62785c093b1834a9526c1fe011c7d23dc6808c045af32e0999dc"} Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.517540 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.520995 4907 scope.go:117] "RemoveContainer" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.528462 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.528494 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.528510 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdxcj\" (UniqueName: \"kubernetes.io/projected/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec-kube-api-access-cdxcj\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.534665 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.539629 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.551213 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.554228 4907 scope.go:117] "RemoveContainer" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.558636 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.568561 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.580495 4907 scope.go:117] "RemoveContainer" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.582336 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.593470 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.600935 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.602274 4907 scope.go:117] "RemoveContainer" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.602756 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c\": container with ID starting with 79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c not found: ID does not exist" containerID="79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.602819 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c"} err="failed to get container status \"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c\": rpc error: code = NotFound desc = could not find container \"79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c\": container with ID starting with 79914c8007c9b10b224f5d4d1fcd270f17078ab7c002b89d3dab089884af5d6c not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.602846 4907 scope.go:117] "RemoveContainer" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.603242 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5\": container with ID starting with 16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5 not found: ID does not exist" containerID="16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603267 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5"} err="failed to get container status \"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5\": rpc error: code = NotFound desc = could not find container \"16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5\": container with ID starting with 16ef32bcd58d9ca78315416c39fe9eca0c5eee8682ac66cea46ec8f9242f8ee5 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603279 4907 scope.go:117] "RemoveContainer" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.603517 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04\": container with ID starting with 78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04 not found: ID does not exist" containerID="78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603551 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04"} err="failed to get container status \"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04\": rpc error: code = NotFound desc = could not find container \"78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04\": container with ID starting with 78891ea90d5eb95d642931ec25514d47603b8811d971d7d33d8fdc5eb0d3bb04 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.603578 4907 scope.go:117] "RemoveContainer" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.603948 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19\": container with ID starting with 285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19 not found: ID does not exist" containerID="285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.604001 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19"} err="failed to get container status \"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19\": rpc error: code = NotFound desc = could not find container \"285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19\": container with ID starting with 285cdb1df855cbeeda75e9d50bffbe69e5c3f9a5485dd118f47a932fb1d36f19 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.604019 4907 scope.go:117] "RemoveContainer" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.607221 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.612402 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-76bbcdf7d9-j6hw5"] Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.619654 4907 scope.go:117] "RemoveContainer" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.620127 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f\": container with ID starting with 5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f not found: ID does not exist" containerID="5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.620158 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f"} err="failed to get container status \"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f\": rpc error: code = NotFound desc = could not find container \"5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f\": container with ID starting with 5ab7cf17ecb1804f362340163cfceb48e6f2c8900108799d01d1025c485c1e8f not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.620177 4907 scope.go:117] "RemoveContainer" containerID="77b52c0f64f55952d2255d5814b60083fd608735671f1f8507b09ea2e83db4d4" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.635203 4907 scope.go:117] "RemoveContainer" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.652627 4907 scope.go:117] "RemoveContainer" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.669200 4907 scope.go:117] "RemoveContainer" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.669615 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176\": container with ID starting with be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176 not found: ID does not exist" containerID="be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.669644 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176"} err="failed to get container status \"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176\": rpc error: code = NotFound desc = could not find container \"be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176\": container with ID starting with be9d4c930500ca917844d7687055220642421756a075ce078989248af8d81176 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.669665 4907 scope.go:117] "RemoveContainer" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.669996 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25\": container with ID starting with b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25 not found: ID does not exist" containerID="b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.670018 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25"} err="failed to get container status \"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25\": rpc error: code = NotFound desc = could not find container \"b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25\": container with ID starting with b5356b2175cd794b4fc6692eb61775ae05e66fb791f7dbd913a5d208e5880c25 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.670030 4907 scope.go:117] "RemoveContainer" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.689829 4907 scope.go:117] "RemoveContainer" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" Mar 13 14:30:51 crc kubenswrapper[4907]: E0313 14:30:51.690319 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56\": container with ID starting with 8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56 not found: ID does not exist" containerID="8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.690357 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56"} err="failed to get container status \"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56\": rpc error: code = NotFound desc = could not find container \"8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56\": container with ID starting with 8a59419fea27690bc54a6fc859fef55fbe86c693159ac3ff37906574c1369e56 not found: ID does not exist" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.805978 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" path="/var/lib/kubelet/pods/25b5fafe-65c8-48ba-bc34-83442f8ace4c/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.807239 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" path="/var/lib/kubelet/pods/25c1c7e6-a149-4a8e-aead-3ebdc7e56851/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.809162 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" path="/var/lib/kubelet/pods/286c82e8-b74a-49d7-a355-ac074aace10e/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.811838 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" path="/var/lib/kubelet/pods/364a702e-e6c6-42bc-8cee-4ba19ea2f1ec/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.813136 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" path="/var/lib/kubelet/pods/4050d00b-0256-45c5-9dc4-0ab46956405d/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.813828 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" path="/var/lib/kubelet/pods/aa81fa2a-4c27-4e02-9891-5b39348feca9/volumes" Mar 13 14:30:51 crc kubenswrapper[4907]: I0313 14:30:51.815687 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" path="/var/lib/kubelet/pods/e3c34454-315e-4821-ab25-b0f331a0d521/volumes" Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.434931 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.435498 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.435899 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.435928 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.438855 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.440327 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.442082 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:52 crc kubenswrapper[4907]: E0313 14:30:52.442118 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:30:53 crc kubenswrapper[4907]: I0313 14:30:53.551029 4907 generic.go:334] "Generic (PLEG): container finished" podID="001078cc-c01e-4173-9740-4f9ff082c593" containerID="8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e" exitCode=0 Mar 13 14:30:53 crc kubenswrapper[4907]: I0313 14:30:53.551207 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerDied","Data":"8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e"} Mar 13 14:30:53 crc kubenswrapper[4907]: I0313 14:30:53.921246 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064419 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064716 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064850 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.064989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.065109 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.065247 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.065361 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") pod \"001078cc-c01e-4173-9740-4f9ff082c593\" (UID: \"001078cc-c01e-4173-9740-4f9ff082c593\") " Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.071486 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.085541 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh" (OuterVolumeSpecName: "kube-api-access-67wmh") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "kube-api-access-67wmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.109048 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.109989 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.117033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.122270 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config" (OuterVolumeSpecName: "config") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.131629 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "001078cc-c01e-4173-9740-4f9ff082c593" (UID: "001078cc-c01e-4173-9740-4f9ff082c593"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167260 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167304 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67wmh\" (UniqueName: \"kubernetes.io/projected/001078cc-c01e-4173-9740-4f9ff082c593-kube-api-access-67wmh\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167316 4907 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167325 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167335 4907 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167346 4907 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.167355 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/001078cc-c01e-4173-9740-4f9ff082c593-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.565595 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bf587b8d9-pvvbz" event={"ID":"001078cc-c01e-4173-9740-4f9ff082c593","Type":"ContainerDied","Data":"502cbc857cf878844cea8a77f9b3d530353b13301737ac4a045e3f57f1cad50e"} Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.565690 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bf587b8d9-pvvbz" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.565922 4907 scope.go:117] "RemoveContainer" containerID="3e3eb6f5fca9c3693edb3599f1d3dd3d706131518c37c580379168a8a1b7e5e0" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.609148 4907 scope.go:117] "RemoveContainer" containerID="8f41b5a31898b937ed3fab2872e25ea71023ac05d2c81d4552d7e5ae8d96095e" Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.610911 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:30:54 crc kubenswrapper[4907]: I0313 14:30:54.617393 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7bf587b8d9-pvvbz"] Mar 13 14:30:55 crc kubenswrapper[4907]: I0313 14:30:55.792336 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="001078cc-c01e-4173-9740-4f9ff082c593" path="/var/lib/kubelet/pods/001078cc-c01e-4173-9740-4f9ff082c593/volumes" Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893634 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893676 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893686 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893698 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:55 crc kubenswrapper[4907]: E0313 14:30:55.893758 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:31:11.893740019 +0000 UTC m=+1570.793527708 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253471 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253747 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253766 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253772 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253780 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253786 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253791 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253797 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253805 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253811 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253820 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253835 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253841 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253854 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253859 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253871 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253876 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253909 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253915 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253922 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253927 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253938 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253944 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253955 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253960 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253970 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253976 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253985 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.253991 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="mysql-bootstrap" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.253998 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254003 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="setup-container" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254016 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254026 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254032 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254042 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254047 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254055 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254061 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254072 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254077 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254085 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254091 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254102 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254107 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254116 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254121 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254129 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254135 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254142 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254147 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254157 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254163 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254173 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254179 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254187 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254192 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254200 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254206 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254213 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254218 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254228 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254233 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254241 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254247 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254253 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254259 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254265 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254270 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254278 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254284 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254293 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254299 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254307 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254313 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254323 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254328 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254336 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254342 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254457 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254467 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254474 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-metadata" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254482 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c1c7e6-a149-4a8e-aead-3ebdc7e56851" containerName="nova-cell1-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254492 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cacb281b-ee3b-42b6-8754-d07e6ea6aa4b" containerName="memcached" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254500 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254507 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254514 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254523 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="001078cc-c01e-4173-9740-4f9ff082c593" containerName="neutron-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254533 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254540 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f8119b2-e38a-494d-967f-5198b83512c7" containerName="ovn-controller" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254548 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="sg-core" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254556 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc227d33-3bf1-4b5f-8368-8e6b8a512f7d" containerName="nova-api-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254563 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254570 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="364a702e-e6c6-42bc-8cee-4ba19ea2f1ec" containerName="nova-cell0-conductor-conductor" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254579 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4050d00b-0256-45c5-9dc4-0ab46956405d" containerName="nova-scheduler-scheduler" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254586 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254594 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c49286b1-2e9f-4de6-b7de-2e952f3ad607" containerName="kube-state-metrics" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254603 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2e62fa-af35-4f35-8852-362ea97c17e7" containerName="glance-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254611 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ab9689-3ab2-42f2-ad56-beb22e29f8da" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254620 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-notification-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254627 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254634 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="ceilometer-central-agent" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254644 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4411bd2-c555-433a-9015-f623948b1401" containerName="galera" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254653 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254661 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4251ae36-90ad-41ea-915e-862df60f5c07" containerName="glance-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254669 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b5fafe-65c8-48ba-bc34-83442f8ace4c" containerName="barbican-worker-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254677 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2f62d4-0584-4cc3-81ed-e067d8db23b1" containerName="barbican-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254685 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a738a66-54bc-4f5e-86d5-6e3004d8f265" containerName="keystone-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254694 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="65531461-cea2-4b2c-a9b6-8fd9e9bffb27" containerName="barbican-keystone-listener-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254702 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c34454-315e-4821-ab25-b0f331a0d521" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254712 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="41d7a0d1-a3ce-45f6-a8b7-fcf9de7f7a76" containerName="cinder-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254721 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3ee069a-41fb-4cab-b650-9bb9c03ac271" containerName="placement-api" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254728 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa81fa2a-4c27-4e02-9891-5b39348feca9" containerName="proxy-httpd" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254736 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e41272f-6e7b-47a7-825e-3571f6a1fd07" containerName="nova-metadata-log" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254744 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="286c82e8-b74a-49d7-a355-ac074aace10e" containerName="rabbitmq" Mar 13 14:30:56 crc kubenswrapper[4907]: E0313 14:30:56.254867 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.254874 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.255006 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="516d1499-3461-4af5-b426-9ae4a711a468" containerName="mariadb-account-create-update" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.255650 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.274928 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.400961 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.401031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.401189 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503152 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503567 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.503700 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.528768 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"redhat-operators-5przm\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:56 crc kubenswrapper[4907]: I0313 14:30:56.576319 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.071761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.434502 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.434964 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.435334 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.435406 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.435633 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.437290 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.438597 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:30:57 crc kubenswrapper[4907]: E0313 14:30:57.438670 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.593644 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" exitCode=0 Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.593714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00"} Mar 13 14:30:57 crc kubenswrapper[4907]: I0313 14:30:57.593758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerStarted","Data":"420aae8884cc744f6583795347d67462ea3274678bbdefc76fe718c1b6545b81"} Mar 13 14:30:58 crc kubenswrapper[4907]: I0313 14:30:58.608059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerStarted","Data":"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04"} Mar 13 14:30:59 crc kubenswrapper[4907]: I0313 14:30:59.619952 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" exitCode=0 Mar 13 14:30:59 crc kubenswrapper[4907]: I0313 14:30:59.620004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04"} Mar 13 14:31:00 crc kubenswrapper[4907]: I0313 14:31:00.632764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerStarted","Data":"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf"} Mar 13 14:31:00 crc kubenswrapper[4907]: I0313 14:31:00.656609 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5przm" podStartSLOduration=2.051220609 podStartE2EDuration="4.656585792s" podCreationTimestamp="2026-03-13 14:30:56 +0000 UTC" firstStartedPulling="2026-03-13 14:30:57.5980306 +0000 UTC m=+1556.497818289" lastFinishedPulling="2026-03-13 14:31:00.203395743 +0000 UTC m=+1559.103183472" observedRunningTime="2026-03-13 14:31:00.652823671 +0000 UTC m=+1559.552611370" watchObservedRunningTime="2026-03-13 14:31:00.656585792 +0000 UTC m=+1559.556373491" Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.433422 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.434184 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.434741 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.434773 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.436529 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.437959 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.439419 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:02 crc kubenswrapper[4907]: E0313 14:31:02.439473 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:06 crc kubenswrapper[4907]: I0313 14:31:06.576792 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:06 crc kubenswrapper[4907]: I0313 14:31:06.577154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.434220 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.435462 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.435828 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.435970 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.436064 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.438070 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.440463 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:07 crc kubenswrapper[4907]: E0313 14:31:07.440525 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:07 crc kubenswrapper[4907]: I0313 14:31:07.621342 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5przm" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" probeResult="failure" output=< Mar 13 14:31:07 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 14:31:07 crc kubenswrapper[4907]: > Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.939458 4907 projected.go:288] Couldn't get configMap openstack/swift-storage-config-data: configmap "swift-storage-config-data" not found Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940087 4907 projected.go:263] Couldn't get secret openstack/swift-conf: secret "swift-conf" not found Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940101 4907 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940112 4907 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:31:11 crc kubenswrapper[4907]: E0313 14:31:11.940223 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift podName:14f05f31-8185-43b0-be69-bcf8d5388ea5 nodeName:}" failed. No retries permitted until 2026-03-13 14:31:43.940204977 +0000 UTC m=+1602.839992666 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift") pod "swift-storage-0" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5") : [configmap "swift-storage-config-data" not found, secret "swift-conf" not found, configmap "swift-ring-files" not found] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.435433 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.436583 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.436627 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.437131 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.437165 4907 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.438263 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.440493 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 14:31:12 crc kubenswrapper[4907]: E0313 14:31:12.440528 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-vp6f5" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.540830 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650746 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650800 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650831 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650897 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.650978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.651000 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") pod \"14f05f31-8185-43b0-be69-bcf8d5388ea5\" (UID: \"14f05f31-8185-43b0-be69-bcf8d5388ea5\") " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.651547 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock" (OuterVolumeSpecName: "lock") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.652103 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache" (OuterVolumeSpecName: "cache") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.656306 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87" (OuterVolumeSpecName: "kube-api-access-7ps87") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "kube-api-access-7ps87". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.656489 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "swift") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.656995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753300 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ps87\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-kube-api-access-7ps87\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753409 4907 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/14f05f31-8185-43b0-be69-bcf8d5388ea5-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753430 4907 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-lock\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753503 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.753523 4907 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/14f05f31-8185-43b0-be69-bcf8d5388ea5-cache\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.769483 4907 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770624 4907 generic.go:334] "Generic (PLEG): container finished" podID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" exitCode=137 Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770662 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea"} Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"14f05f31-8185-43b0-be69-bcf8d5388ea5","Type":"ContainerDied","Data":"8ab726bec0bb062d6ac5d70e873cfd78800b9921e7163b9cb0ca74467cc3c7f4"} Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770708 4907 scope.go:117] "RemoveContainer" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.770757 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.854381 4907 scope.go:117] "RemoveContainer" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.855368 4907 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.879977 4907 scope.go:117] "RemoveContainer" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.913744 4907 scope.go:117] "RemoveContainer" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.943539 4907 scope.go:117] "RemoveContainer" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.972200 4907 scope.go:117] "RemoveContainer" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.974841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14f05f31-8185-43b0-be69-bcf8d5388ea5" (UID: "14f05f31-8185-43b0-be69-bcf8d5388ea5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:31:12 crc kubenswrapper[4907]: I0313 14:31:12.997393 4907 scope.go:117] "RemoveContainer" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.021719 4907 scope.go:117] "RemoveContainer" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.048177 4907 scope.go:117] "RemoveContainer" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.061529 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14f05f31-8185-43b0-be69-bcf8d5388ea5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.068156 4907 scope.go:117] "RemoveContainer" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.104748 4907 scope.go:117] "RemoveContainer" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.138490 4907 scope.go:117] "RemoveContainer" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.145524 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.152819 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.193453 4907 scope.go:117] "RemoveContainer" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.219025 4907 scope.go:117] "RemoveContainer" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.274519 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vp6f5_c1b70392-1240-40d9-8128-e7abe29c8398/ovs-vswitchd/0.log" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.282419 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.287127 4907 scope.go:117] "RemoveContainer" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.308506 4907 scope.go:117] "RemoveContainer" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309048 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea\": container with ID starting with 0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea not found: ID does not exist" containerID="0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309090 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea"} err="failed to get container status \"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea\": rpc error: code = NotFound desc = could not find container \"0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea\": container with ID starting with 0696dc2db616f4616e43dd171ba757d11bce09169b35b516964d1af58bad39ea not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309116 4907 scope.go:117] "RemoveContainer" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309374 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411\": container with ID starting with 1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411 not found: ID does not exist" containerID="1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309394 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411"} err="failed to get container status \"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411\": rpc error: code = NotFound desc = could not find container \"1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411\": container with ID starting with 1e5ba68b21b276693f03089f51ca3955dd88adcaacae52727d03a999b543d411 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309406 4907 scope.go:117] "RemoveContainer" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309615 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2\": container with ID starting with 8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2 not found: ID does not exist" containerID="8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309655 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2"} err="failed to get container status \"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2\": rpc error: code = NotFound desc = could not find container \"8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2\": container with ID starting with 8765bf4cde625366ddd57fa87118a09b18b3fbbf1d2ded31c12344e9c08dadb2 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309668 4907 scope.go:117] "RemoveContainer" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.309817 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14\": container with ID starting with 6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14 not found: ID does not exist" containerID="6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309835 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14"} err="failed to get container status \"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14\": rpc error: code = NotFound desc = could not find container \"6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14\": container with ID starting with 6f5cf167a5953dd1bdb0f436b238215d64975f16e188a1dacd546c8d2adfba14 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.309847 4907 scope.go:117] "RemoveContainer" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310048 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4\": container with ID starting with c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4 not found: ID does not exist" containerID="c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310068 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4"} err="failed to get container status \"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4\": rpc error: code = NotFound desc = could not find container \"c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4\": container with ID starting with c1b4fe0c395990393cfbace1600ea9872fde43535722ed738332ba0b26f178a4 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310078 4907 scope.go:117] "RemoveContainer" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310210 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94\": container with ID starting with 5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94 not found: ID does not exist" containerID="5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310228 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94"} err="failed to get container status \"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94\": rpc error: code = NotFound desc = could not find container \"5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94\": container with ID starting with 5a6b392bf5f4cbd99ba2fc751327322f8960f4c73aba15cb9a501d4c9c3edd94 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310238 4907 scope.go:117] "RemoveContainer" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310402 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7\": container with ID starting with c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7 not found: ID does not exist" containerID="c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310421 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7"} err="failed to get container status \"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7\": rpc error: code = NotFound desc = could not find container \"c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7\": container with ID starting with c81c73db839bc006e1b7fdce225dd05a9a50870ea60ca389ea146e1e69d195b7 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310432 4907 scope.go:117] "RemoveContainer" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310585 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f\": container with ID starting with dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f not found: ID does not exist" containerID="dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310602 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f"} err="failed to get container status \"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f\": rpc error: code = NotFound desc = could not find container \"dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f\": container with ID starting with dbcb2b472a15fbcc053532418c2f7a654827914050e539781030bc0749d6218f not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310614 4907 scope.go:117] "RemoveContainer" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310754 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808\": container with ID starting with 7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808 not found: ID does not exist" containerID="7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310774 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808"} err="failed to get container status \"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808\": rpc error: code = NotFound desc = could not find container \"7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808\": container with ID starting with 7a94b1f20e2cb964aaf2fe5c9f23f0cd193291a14edd5411717f7cad81ed9808 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310787 4907 scope.go:117] "RemoveContainer" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.310942 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80\": container with ID starting with 2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80 not found: ID does not exist" containerID="2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310966 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80"} err="failed to get container status \"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80\": rpc error: code = NotFound desc = could not find container \"2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80\": container with ID starting with 2bc2275a73f47c1a27db775d1cc1c3ae9c20c01f9ddad7a35bf652467b66ea80 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.310978 4907 scope.go:117] "RemoveContainer" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311155 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a\": container with ID starting with b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a not found: ID does not exist" containerID="b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311172 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a"} err="failed to get container status \"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a\": rpc error: code = NotFound desc = could not find container \"b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a\": container with ID starting with b7f60cadabc1fd2d193073423cce7296e305ff6325001c7ec9823dd267a1787a not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311184 4907 scope.go:117] "RemoveContainer" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311321 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410\": container with ID starting with c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410 not found: ID does not exist" containerID="c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311339 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410"} err="failed to get container status \"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410\": rpc error: code = NotFound desc = could not find container \"c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410\": container with ID starting with c7370f4c24ed7b82c59bc2c1e3ae152c65420b5dd6227a66eb64b5b83832e410 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311350 4907 scope.go:117] "RemoveContainer" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311519 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e\": container with ID starting with f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e not found: ID does not exist" containerID="f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311537 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e"} err="failed to get container status \"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e\": rpc error: code = NotFound desc = could not find container \"f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e\": container with ID starting with f8bc1f3dd2242db8254b664f00b3789940491fa59b38a7fc033c225b233b776e not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311551 4907 scope.go:117] "RemoveContainer" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.311754 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566\": container with ID starting with 9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566 not found: ID does not exist" containerID="9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311863 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566"} err="failed to get container status \"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566\": rpc error: code = NotFound desc = could not find container \"9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566\": container with ID starting with 9b6f16a5610e8f2c08a40a29bfba0c00e1ff883cb29b5290604713aad1dab566 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.311889 4907 scope.go:117] "RemoveContainer" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.312219 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091\": container with ID starting with 8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091 not found: ID does not exist" containerID="8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.312269 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091"} err="failed to get container status \"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091\": rpc error: code = NotFound desc = could not find container \"8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091\": container with ID starting with 8168297a51ad087d7375035164e54e4a0d24e6fd56cf85e922e11d80f7f3d091 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365734 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log" (OuterVolumeSpecName: "var-log") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365775 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365792 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib" (OuterVolumeSpecName: "var-lib") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365824 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365866 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.365900 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") pod \"c1b70392-1240-40d9-8128-e7abe29c8398\" (UID: \"c1b70392-1240-40d9-8128-e7abe29c8398\") " Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366054 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366017 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run" (OuterVolumeSpecName: "var-run") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366269 4907 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-lib\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366285 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366297 4907 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-etc-ovs\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.366309 4907 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/c1b70392-1240-40d9-8128-e7abe29c8398-var-log\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.367198 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts" (OuterVolumeSpecName: "scripts") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.373463 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj" (OuterVolumeSpecName: "kube-api-access-4qvcj") pod "c1b70392-1240-40d9-8128-e7abe29c8398" (UID: "c1b70392-1240-40d9-8128-e7abe29c8398"). InnerVolumeSpecName "kube-api-access-4qvcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.467575 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qvcj\" (UniqueName: \"kubernetes.io/projected/c1b70392-1240-40d9-8128-e7abe29c8398-kube-api-access-4qvcj\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.467615 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1b70392-1240-40d9-8128-e7abe29c8398-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.784971 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vp6f5_c1b70392-1240-40d9-8128-e7abe29c8398/ovs-vswitchd/0.log" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.786120 4907 generic.go:334] "Generic (PLEG): container finished" podID="c1b70392-1240-40d9-8128-e7abe29c8398" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" exitCode=137 Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.786228 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vp6f5" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.789979 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" path="/var/lib/kubelet/pods/14f05f31-8185-43b0-be69-bcf8d5388ea5/volumes" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.792520 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f"} Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.792581 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vp6f5" event={"ID":"c1b70392-1240-40d9-8128-e7abe29c8398","Type":"ContainerDied","Data":"99abe9d38052154162343b52b5e60bd6560b49e824682c932c7a2a64adbb5d61"} Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.792607 4907 scope.go:117] "RemoveContainer" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.810845 4907 scope.go:117] "RemoveContainer" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.833058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.841067 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-vp6f5"] Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.843091 4907 scope.go:117] "RemoveContainer" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.873460 4907 scope.go:117] "RemoveContainer" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.874307 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f\": container with ID starting with 265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f not found: ID does not exist" containerID="265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874350 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f"} err="failed to get container status \"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f\": rpc error: code = NotFound desc = could not find container \"265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f\": container with ID starting with 265b6b6fe5f95195cdc86c65456998ad327fc7ff3a6641300cd6ac00798eff4f not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874379 4907 scope.go:117] "RemoveContainer" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.874706 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574\": container with ID starting with 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 not found: ID does not exist" containerID="6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874729 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574"} err="failed to get container status \"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574\": rpc error: code = NotFound desc = could not find container \"6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574\": container with ID starting with 6b0b25079cc099dabe922ac89fa41ae76d2f01797aac9f273b042e242459c574 not found: ID does not exist" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.874747 4907 scope.go:117] "RemoveContainer" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" Mar 13 14:31:13 crc kubenswrapper[4907]: E0313 14:31:13.875229 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31\": container with ID starting with 4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31 not found: ID does not exist" containerID="4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31" Mar 13 14:31:13 crc kubenswrapper[4907]: I0313 14:31:13.875263 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31"} err="failed to get container status \"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31\": rpc error: code = NotFound desc = could not find container \"4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31\": container with ID starting with 4cf539e8b8cde03127504ad0dcb18c074f6f0c3dab53805c59e991c594c32a31 not found: ID does not exist" Mar 13 14:31:15 crc kubenswrapper[4907]: I0313 14:31:15.791758 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" path="/var/lib/kubelet/pods/c1b70392-1240-40d9-8128-e7abe29c8398/volumes" Mar 13 14:31:16 crc kubenswrapper[4907]: I0313 14:31:16.616847 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:16 crc kubenswrapper[4907]: I0313 14:31:16.660987 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:16 crc kubenswrapper[4907]: I0313 14:31:16.845009 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:31:17 crc kubenswrapper[4907]: I0313 14:31:17.824407 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5przm" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" containerID="cri-o://b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" gracePeriod=2 Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.042311 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.042867 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.321682 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.432956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") pod \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.433009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") pod \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.433086 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") pod \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\" (UID: \"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa\") " Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.434384 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities" (OuterVolumeSpecName: "utilities") pod "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" (UID: "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.443375 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd" (OuterVolumeSpecName: "kube-api-access-g8dvd") pod "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" (UID: "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa"). InnerVolumeSpecName "kube-api-access-g8dvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.534679 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.535000 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8dvd\" (UniqueName: \"kubernetes.io/projected/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-kube-api-access-g8dvd\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.583033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" (UID: "1e43e4fe-9084-48bb-a931-33e2fc5ecaaa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.636515 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840640 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" exitCode=0 Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840701 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf"} Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5przm" event={"ID":"1e43e4fe-9084-48bb-a931-33e2fc5ecaaa","Type":"ContainerDied","Data":"420aae8884cc744f6583795347d67462ea3274678bbdefc76fe718c1b6545b81"} Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840742 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5przm" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.840771 4907 scope.go:117] "RemoveContainer" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.867334 4907 scope.go:117] "RemoveContainer" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.891946 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.900009 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5przm"] Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.919539 4907 scope.go:117] "RemoveContainer" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.945823 4907 scope.go:117] "RemoveContainer" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" Mar 13 14:31:18 crc kubenswrapper[4907]: E0313 14:31:18.948366 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf\": container with ID starting with b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf not found: ID does not exist" containerID="b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.948460 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf"} err="failed to get container status \"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf\": rpc error: code = NotFound desc = could not find container \"b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf\": container with ID starting with b4fc4a9c98855f93f56a77779ad655ec33841f4a3e7acba65768c38b4a3718bf not found: ID does not exist" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.948562 4907 scope.go:117] "RemoveContainer" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" Mar 13 14:31:18 crc kubenswrapper[4907]: E0313 14:31:18.951127 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04\": container with ID starting with 057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04 not found: ID does not exist" containerID="057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.951250 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04"} err="failed to get container status \"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04\": rpc error: code = NotFound desc = could not find container \"057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04\": container with ID starting with 057b7d13b67e10cac7b4063457b964d60358da6d919d6ef053cd4aaf19089d04 not found: ID does not exist" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.951295 4907 scope.go:117] "RemoveContainer" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" Mar 13 14:31:18 crc kubenswrapper[4907]: E0313 14:31:18.957300 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00\": container with ID starting with 4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00 not found: ID does not exist" containerID="4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00" Mar 13 14:31:18 crc kubenswrapper[4907]: I0313 14:31:18.957359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00"} err="failed to get container status \"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00\": rpc error: code = NotFound desc = could not find container \"4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00\": container with ID starting with 4b6734fa58a17ab6193b32c4603463a610dfd0ffb321199337a7a16643dced00 not found: ID does not exist" Mar 13 14:31:19 crc kubenswrapper[4907]: I0313 14:31:19.793769 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" path="/var/lib/kubelet/pods/1e43e4fe-9084-48bb-a931-33e2fc5ecaaa/volumes" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.528160 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529468 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529487 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529668 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529682 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529690 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-content" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529699 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-content" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529717 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529725 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529742 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529749 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529759 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529767 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529781 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529789 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529798 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529806 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529820 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529830 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529841 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529848 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529858 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529866 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529893 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529901 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529912 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529933 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529941 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529955 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server-init" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529962 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server-init" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529973 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529980 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.529990 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.529998 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530006 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530013 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530024 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530033 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530047 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-utilities" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530055 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="extract-utilities" Mar 13 14:31:30 crc kubenswrapper[4907]: E0313 14:31:30.530064 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530070 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530248 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530265 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-expirer" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530277 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530291 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="swift-recon-cron" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530303 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530311 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-reaper" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530320 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530328 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-updater" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530341 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="rsync" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530350 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovsdb-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530361 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530371 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530381 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e43e4fe-9084-48bb-a931-33e2fc5ecaaa" containerName="registry-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530389 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="account-replicator" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530401 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="object-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530412 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-server" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530424 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1b70392-1240-40d9-8128-e7abe29c8398" containerName="ovs-vswitchd" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.530433 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f05f31-8185-43b0-be69-bcf8d5388ea5" containerName="container-auditor" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.531597 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.545362 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.702171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.702231 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.702494 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804392 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804527 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.804867 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.805289 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.824677 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"certified-operators-2dpx8\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:30 crc kubenswrapper[4907]: I0313 14:31:30.855614 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.349444 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.967708 4907 generic.go:334] "Generic (PLEG): container finished" podID="d9c047f6-207b-451e-acb0-e909774dc44e" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" exitCode=0 Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.967802 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961"} Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.968076 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerStarted","Data":"86437821014ef87426f69aad718b9aa1d504d40345fbe79e3129c90fbf55b0d2"} Mar 13 14:31:31 crc kubenswrapper[4907]: I0313 14:31:31.969986 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:31:32 crc kubenswrapper[4907]: I0313 14:31:32.979175 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerStarted","Data":"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e"} Mar 13 14:31:33 crc kubenswrapper[4907]: I0313 14:31:33.990978 4907 generic.go:334] "Generic (PLEG): container finished" podID="d9c047f6-207b-451e-acb0-e909774dc44e" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" exitCode=0 Mar 13 14:31:33 crc kubenswrapper[4907]: I0313 14:31:33.991032 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e"} Mar 13 14:31:35 crc kubenswrapper[4907]: I0313 14:31:35.002265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerStarted","Data":"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c"} Mar 13 14:31:35 crc kubenswrapper[4907]: I0313 14:31:35.020458 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2dpx8" podStartSLOduration=2.458874806 podStartE2EDuration="5.020437561s" podCreationTimestamp="2026-03-13 14:31:30 +0000 UTC" firstStartedPulling="2026-03-13 14:31:31.969588264 +0000 UTC m=+1590.869375983" lastFinishedPulling="2026-03-13 14:31:34.531151049 +0000 UTC m=+1593.430938738" observedRunningTime="2026-03-13 14:31:35.019114285 +0000 UTC m=+1593.918901984" watchObservedRunningTime="2026-03-13 14:31:35.020437561 +0000 UTC m=+1593.920225250" Mar 13 14:31:40 crc kubenswrapper[4907]: I0313 14:31:40.855982 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:40 crc kubenswrapper[4907]: I0313 14:31:40.856308 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:40 crc kubenswrapper[4907]: I0313 14:31:40.901049 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:41 crc kubenswrapper[4907]: I0313 14:31:41.104754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:41 crc kubenswrapper[4907]: I0313 14:31:41.161658 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.076987 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2dpx8" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" containerID="cri-o://e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" gracePeriod=2 Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.437747 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.589975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") pod \"d9c047f6-207b-451e-acb0-e909774dc44e\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.590026 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") pod \"d9c047f6-207b-451e-acb0-e909774dc44e\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.590050 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") pod \"d9c047f6-207b-451e-acb0-e909774dc44e\" (UID: \"d9c047f6-207b-451e-acb0-e909774dc44e\") " Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.591143 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities" (OuterVolumeSpecName: "utilities") pod "d9c047f6-207b-451e-acb0-e909774dc44e" (UID: "d9c047f6-207b-451e-acb0-e909774dc44e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.595763 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l" (OuterVolumeSpecName: "kube-api-access-ttd6l") pod "d9c047f6-207b-451e-acb0-e909774dc44e" (UID: "d9c047f6-207b-451e-acb0-e909774dc44e"). InnerVolumeSpecName "kube-api-access-ttd6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.692036 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttd6l\" (UniqueName: \"kubernetes.io/projected/d9c047f6-207b-451e-acb0-e909774dc44e-kube-api-access-ttd6l\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:43 crc kubenswrapper[4907]: I0313 14:31:43.692068 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.088978 4907 generic.go:334] "Generic (PLEG): container finished" podID="d9c047f6-207b-451e-acb0-e909774dc44e" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" exitCode=0 Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c"} Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089110 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dpx8" event={"ID":"d9c047f6-207b-451e-acb0-e909774dc44e","Type":"ContainerDied","Data":"86437821014ef87426f69aad718b9aa1d504d40345fbe79e3129c90fbf55b0d2"} Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089138 4907 scope.go:117] "RemoveContainer" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.089044 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dpx8" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.109195 4907 scope.go:117] "RemoveContainer" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.130990 4907 scope.go:117] "RemoveContainer" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.172987 4907 scope.go:117] "RemoveContainer" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" Mar 13 14:31:44 crc kubenswrapper[4907]: E0313 14:31:44.173543 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c\": container with ID starting with e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c not found: ID does not exist" containerID="e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173582 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c"} err="failed to get container status \"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c\": rpc error: code = NotFound desc = could not find container \"e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c\": container with ID starting with e745fb2cd1882130e47435a6e6cfa910c9eb1a8493a9ef4cf1add5adbf625f8c not found: ID does not exist" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173605 4907 scope.go:117] "RemoveContainer" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" Mar 13 14:31:44 crc kubenswrapper[4907]: E0313 14:31:44.173894 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e\": container with ID starting with a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e not found: ID does not exist" containerID="a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173927 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e"} err="failed to get container status \"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e\": rpc error: code = NotFound desc = could not find container \"a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e\": container with ID starting with a97db1b0d0bfe03ca4dcb913a70fc3028d28f841560e547d602b7d23a4710e8e not found: ID does not exist" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.173947 4907 scope.go:117] "RemoveContainer" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" Mar 13 14:31:44 crc kubenswrapper[4907]: E0313 14:31:44.174228 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961\": container with ID starting with 764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961 not found: ID does not exist" containerID="764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.174250 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961"} err="failed to get container status \"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961\": rpc error: code = NotFound desc = could not find container \"764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961\": container with ID starting with 764acaf87f75bed17bfb66456c3fd4c1cb85fa1c820674eb12e9306b1a048961 not found: ID does not exist" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.270264 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d9c047f6-207b-451e-acb0-e909774dc44e" (UID: "d9c047f6-207b-451e-acb0-e909774dc44e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.301509 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d9c047f6-207b-451e-acb0-e909774dc44e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.423525 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:44 crc kubenswrapper[4907]: I0313 14:31:44.431048 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2dpx8"] Mar 13 14:31:45 crc kubenswrapper[4907]: I0313 14:31:45.792844 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" path="/var/lib/kubelet/pods/d9c047f6-207b-451e-acb0-e909774dc44e/volumes" Mar 13 14:31:48 crc kubenswrapper[4907]: I0313 14:31:48.041869 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:31:48 crc kubenswrapper[4907]: I0313 14:31:48.042746 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.145460 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:32:00 crc kubenswrapper[4907]: E0313 14:32:00.146416 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" Mar 13 14:32:00 crc kubenswrapper[4907]: E0313 14:32:00.146457 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-content" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146465 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-content" Mar 13 14:32:00 crc kubenswrapper[4907]: E0313 14:32:00.146478 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-utilities" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146486 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="extract-utilities" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.146638 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c047f6-207b-451e-acb0-e909774dc44e" containerName="registry-server" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.147235 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.152124 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.152683 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.155632 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.158646 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.312793 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"auto-csr-approver-29556872-srgh9\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.413726 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"auto-csr-approver-29556872-srgh9\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.435165 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"auto-csr-approver-29556872-srgh9\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.468560 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.885270 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:32:00 crc kubenswrapper[4907]: I0313 14:32:00.992891 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556872-srgh9" event={"ID":"32297c61-6461-4357-832f-9de703d88000","Type":"ContainerStarted","Data":"4b360592a61dc96e94c5095ba7d5510824692c1ccb09aa93abcdd78a8df14367"} Mar 13 14:32:03 crc kubenswrapper[4907]: I0313 14:32:03.011514 4907 generic.go:334] "Generic (PLEG): container finished" podID="32297c61-6461-4357-832f-9de703d88000" containerID="8eea35f2339f97a9cd81df041bdf950f47b0d8ec3a69dea570886bdde6ad88a7" exitCode=0 Mar 13 14:32:03 crc kubenswrapper[4907]: I0313 14:32:03.011622 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556872-srgh9" event={"ID":"32297c61-6461-4357-832f-9de703d88000","Type":"ContainerDied","Data":"8eea35f2339f97a9cd81df041bdf950f47b0d8ec3a69dea570886bdde6ad88a7"} Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.324175 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.475253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") pod \"32297c61-6461-4357-832f-9de703d88000\" (UID: \"32297c61-6461-4357-832f-9de703d88000\") " Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.481553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt" (OuterVolumeSpecName: "kube-api-access-wfpwt") pod "32297c61-6461-4357-832f-9de703d88000" (UID: "32297c61-6461-4357-832f-9de703d88000"). InnerVolumeSpecName "kube-api-access-wfpwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:32:04 crc kubenswrapper[4907]: I0313 14:32:04.577556 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfpwt\" (UniqueName: \"kubernetes.io/projected/32297c61-6461-4357-832f-9de703d88000-kube-api-access-wfpwt\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.026602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556872-srgh9" event={"ID":"32297c61-6461-4357-832f-9de703d88000","Type":"ContainerDied","Data":"4b360592a61dc96e94c5095ba7d5510824692c1ccb09aa93abcdd78a8df14367"} Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.026642 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b360592a61dc96e94c5095ba7d5510824692c1ccb09aa93abcdd78a8df14367" Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.026672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556872-srgh9" Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.393785 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.400003 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556866-fwg48"] Mar 13 14:32:05 crc kubenswrapper[4907]: I0313 14:32:05.793311 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d930b98-c1af-4e47-a663-86afa9484856" path="/var/lib/kubelet/pods/7d930b98-c1af-4e47-a663-86afa9484856/volumes" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.041437 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042034 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042077 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042715 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:32:18 crc kubenswrapper[4907]: I0313 14:32:18.042770 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" gracePeriod=600 Mar 13 14:32:18 crc kubenswrapper[4907]: E0313 14:32:18.171104 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.142250 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" exitCode=0 Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.142295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2"} Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.142332 4907 scope.go:117] "RemoveContainer" containerID="1f79a88cd4431d2b6772ef1dc1fb18ce20d965a3a936a7157e03bc28ca239c57" Mar 13 14:32:19 crc kubenswrapper[4907]: I0313 14:32:19.143122 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:32:19 crc kubenswrapper[4907]: E0313 14:32:19.143512 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.539643 4907 scope.go:117] "RemoveContainer" containerID="93ab187126d08591f8a6823b5ca150c5522a7db6aea7e1ca8cc868ee0d426800" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.562639 4907 scope.go:117] "RemoveContainer" containerID="37d0b4abf880f9988e205a5d09debf0e634cb45e2049ee542b43a256eccbec25" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.624463 4907 scope.go:117] "RemoveContainer" containerID="65475a4b4006963b7fec1fa34278a44b7bf08cbcf9c088a764d7d29c870c630c" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.648198 4907 scope.go:117] "RemoveContainer" containerID="b2e917cc4c559e06be2ed2a503af651b2cead394bb06409f6e1c888788df817a" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.666477 4907 scope.go:117] "RemoveContainer" containerID="d5c1cbc11101f46dd814717af56c35f8ffa0c1cd1640b4a7f3a2d4236359a48e" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.691329 4907 scope.go:117] "RemoveContainer" containerID="ae6341ae2f61b11a2ad0c4f86bf51fc609fce628a9af3e20ce77aa5f1d917a69" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.712484 4907 scope.go:117] "RemoveContainer" containerID="1953abb8d3f9054081c314893bde9d173fee7430bf61ccd6ae11b908a7002729" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.743136 4907 scope.go:117] "RemoveContainer" containerID="e6d88705fdc67e86fd0214f4c6bbf1c37e140065fdd712f700392f915f879795" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.762120 4907 scope.go:117] "RemoveContainer" containerID="3d70e191441a97faf79353905c86d9af8e9fd167c63431a9f984ad9d2a36445e" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.778796 4907 scope.go:117] "RemoveContainer" containerID="b5e68fc91f35838327b362dd7ecd6d2f5ee1173bc7ae4fafcb541ac0c96b44fa" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.799167 4907 scope.go:117] "RemoveContainer" containerID="c688bfe6d6ea48e683001f8125ee585ba89819e96749b44bbe5f071b99df1226" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.821586 4907 scope.go:117] "RemoveContainer" containerID="1a454f90b2465a04739b60faa1ea54a4c62cae0b7a6742d83f6b477594d8a8a9" Mar 13 14:32:22 crc kubenswrapper[4907]: I0313 14:32:22.843779 4907 scope.go:117] "RemoveContainer" containerID="3c90645adf755c84a8db83dd2e03403f66a250afbd541346b9261208e83c56ff" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.883195 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:29 crc kubenswrapper[4907]: E0313 14:32:29.885382 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32297c61-6461-4357-832f-9de703d88000" containerName="oc" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.885405 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="32297c61-6461-4357-832f-9de703d88000" containerName="oc" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.885607 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="32297c61-6461-4357-832f-9de703d88000" containerName="oc" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.886870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.898074 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.926453 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.926527 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:29 crc kubenswrapper[4907]: I0313 14:32:29.926544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.027747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.027790 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.027855 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.028354 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.028444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.049006 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"community-operators-4rn6k\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.212268 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:30 crc kubenswrapper[4907]: I0313 14:32:30.690100 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:31 crc kubenswrapper[4907]: I0313 14:32:31.262994 4907 generic.go:334] "Generic (PLEG): container finished" podID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" exitCode=0 Mar 13 14:32:31 crc kubenswrapper[4907]: I0313 14:32:31.263178 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67"} Mar 13 14:32:31 crc kubenswrapper[4907]: I0313 14:32:31.263397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerStarted","Data":"13c2f49681df8442c517f0c438f3ee426387fdb04f4ee6868a9b1082ae36005d"} Mar 13 14:32:32 crc kubenswrapper[4907]: I0313 14:32:32.276161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerStarted","Data":"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd"} Mar 13 14:32:32 crc kubenswrapper[4907]: I0313 14:32:32.782898 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:32:32 crc kubenswrapper[4907]: E0313 14:32:32.783191 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:32:33 crc kubenswrapper[4907]: I0313 14:32:33.287685 4907 generic.go:334] "Generic (PLEG): container finished" podID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" exitCode=0 Mar 13 14:32:33 crc kubenswrapper[4907]: I0313 14:32:33.287779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd"} Mar 13 14:32:34 crc kubenswrapper[4907]: I0313 14:32:34.299122 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerStarted","Data":"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612"} Mar 13 14:32:34 crc kubenswrapper[4907]: I0313 14:32:34.319554 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4rn6k" podStartSLOduration=2.747760435 podStartE2EDuration="5.319532045s" podCreationTimestamp="2026-03-13 14:32:29 +0000 UTC" firstStartedPulling="2026-03-13 14:32:31.26784225 +0000 UTC m=+1650.167629949" lastFinishedPulling="2026-03-13 14:32:33.83961387 +0000 UTC m=+1652.739401559" observedRunningTime="2026-03-13 14:32:34.318675883 +0000 UTC m=+1653.218463572" watchObservedRunningTime="2026-03-13 14:32:34.319532045 +0000 UTC m=+1653.219319734" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.212820 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.213400 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.271814 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.423174 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:40 crc kubenswrapper[4907]: I0313 14:32:40.512311 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:42 crc kubenswrapper[4907]: I0313 14:32:42.373864 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4rn6k" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" containerID="cri-o://aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" gracePeriod=2 Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.336413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386649 4907 generic.go:334] "Generic (PLEG): container finished" podID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" exitCode=0 Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386696 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612"} Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386725 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4rn6k" event={"ID":"bf79c546-2d9a-4eba-bf86-f78d196bc557","Type":"ContainerDied","Data":"13c2f49681df8442c517f0c438f3ee426387fdb04f4ee6868a9b1082ae36005d"} Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386744 4907 scope.go:117] "RemoveContainer" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.386921 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4rn6k" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.420367 4907 scope.go:117] "RemoveContainer" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.428431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") pod \"bf79c546-2d9a-4eba-bf86-f78d196bc557\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.428548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") pod \"bf79c546-2d9a-4eba-bf86-f78d196bc557\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.428579 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") pod \"bf79c546-2d9a-4eba-bf86-f78d196bc557\" (UID: \"bf79c546-2d9a-4eba-bf86-f78d196bc557\") " Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.430258 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities" (OuterVolumeSpecName: "utilities") pod "bf79c546-2d9a-4eba-bf86-f78d196bc557" (UID: "bf79c546-2d9a-4eba-bf86-f78d196bc557"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.438217 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4" (OuterVolumeSpecName: "kube-api-access-cxlz4") pod "bf79c546-2d9a-4eba-bf86-f78d196bc557" (UID: "bf79c546-2d9a-4eba-bf86-f78d196bc557"). InnerVolumeSpecName "kube-api-access-cxlz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.498196 4907 scope.go:117] "RemoveContainer" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.519244 4907 scope.go:117] "RemoveContainer" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" Mar 13 14:32:43 crc kubenswrapper[4907]: E0313 14:32:43.519642 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612\": container with ID starting with aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612 not found: ID does not exist" containerID="aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.519674 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612"} err="failed to get container status \"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612\": rpc error: code = NotFound desc = could not find container \"aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612\": container with ID starting with aaee29c2d401955a8cb458cf5d817e6a9250580d773d1a23710ccd6b61e1d612 not found: ID does not exist" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.519693 4907 scope.go:117] "RemoveContainer" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" Mar 13 14:32:43 crc kubenswrapper[4907]: E0313 14:32:43.520096 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd\": container with ID starting with b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd not found: ID does not exist" containerID="b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.520119 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd"} err="failed to get container status \"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd\": rpc error: code = NotFound desc = could not find container \"b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd\": container with ID starting with b81eb78ae61aa093de15332c1ee04faf01a340ba43eebb6b2242c49cc2235cbd not found: ID does not exist" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.520133 4907 scope.go:117] "RemoveContainer" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" Mar 13 14:32:43 crc kubenswrapper[4907]: E0313 14:32:43.520418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67\": container with ID starting with 4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67 not found: ID does not exist" containerID="4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.520436 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67"} err="failed to get container status \"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67\": rpc error: code = NotFound desc = could not find container \"4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67\": container with ID starting with 4bf9c94aaa9cc7c5c0ba8069c717611a0d87cc031d7ffb209791060330b75f67 not found: ID does not exist" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.529706 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxlz4\" (UniqueName: \"kubernetes.io/projected/bf79c546-2d9a-4eba-bf86-f78d196bc557-kube-api-access-cxlz4\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.529749 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.531950 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf79c546-2d9a-4eba-bf86-f78d196bc557" (UID: "bf79c546-2d9a-4eba-bf86-f78d196bc557"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.631013 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf79c546-2d9a-4eba-bf86-f78d196bc557-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.715986 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.723073 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4rn6k"] Mar 13 14:32:43 crc kubenswrapper[4907]: I0313 14:32:43.797264 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" path="/var/lib/kubelet/pods/bf79c546-2d9a-4eba-bf86-f78d196bc557/volumes" Mar 13 14:32:45 crc kubenswrapper[4907]: I0313 14:32:45.783495 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:32:45 crc kubenswrapper[4907]: E0313 14:32:45.783791 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:00 crc kubenswrapper[4907]: I0313 14:33:00.782149 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:00 crc kubenswrapper[4907]: E0313 14:33:00.783200 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:13 crc kubenswrapper[4907]: I0313 14:33:13.782873 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:13 crc kubenswrapper[4907]: E0313 14:33:13.784904 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.091140 4907 scope.go:117] "RemoveContainer" containerID="8a34f94ac114c7c6f4deeb646d486da44df7c2de5f8d5ce22726005779d1271f" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.128715 4907 scope.go:117] "RemoveContainer" containerID="8f4cb6476a276775b25f72fa46f11e3f550f531236e5cf19234f8b274637da17" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.147224 4907 scope.go:117] "RemoveContainer" containerID="4e3ebec78f62227a1eec084cd7d40d9bb86b61261221bf385169185e602cfad5" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.175111 4907 scope.go:117] "RemoveContainer" containerID="fad76ed01fe4f9787942e4f47ff8ffafe18d1d8c62a2533bd080f0d2d246c933" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.210033 4907 scope.go:117] "RemoveContainer" containerID="3e1e04ee01cb4c593a590da900f137153aed8f7c58309d1a24f19dd94c9af998" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.231657 4907 scope.go:117] "RemoveContainer" containerID="5b6a2afb9b4ad5f576a34ca212d94796e12f98db7c7c3a2d4bb1600fc3d60034" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.260445 4907 scope.go:117] "RemoveContainer" containerID="70df9585a66270db2bec365cac2492109be5b468d9fc1bab776dec0ef7304eaa" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.297041 4907 scope.go:117] "RemoveContainer" containerID="c90c150f12af47e1b50f63c9ddceacb4d1551893cbeff831172fa51757c596cb" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.332855 4907 scope.go:117] "RemoveContainer" containerID="cdd8592044b8b42302c000b60e593bcd7aa7352aa0bfd95a951e13ddb4d1cef1" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.350976 4907 scope.go:117] "RemoveContainer" containerID="f407f7b1e1671cab73872a73a57d80595693b20473de90bda4d6de97abc60cd5" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.379347 4907 scope.go:117] "RemoveContainer" containerID="0a3f47d3069fe894c10b414d1b636502b45151a11f2acdaa45abe3e56ee5a7a1" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.402739 4907 scope.go:117] "RemoveContainer" containerID="46c2ff1e9e66c6fc449295a888dc400f857863d7df33a386b4b1e1cf42f6a3f4" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.427217 4907 scope.go:117] "RemoveContainer" containerID="2ad298467d04021b9b1a343e12a26a79c8a6605114bfe6ee4b9cfd24470e49c4" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.468019 4907 scope.go:117] "RemoveContainer" containerID="71b8e65c6bc18f48a724ed12b9dcf478104abf3f0beba637548e1962f43f12bb" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.485560 4907 scope.go:117] "RemoveContainer" containerID="766d1e6536dea7ad49da9b30346790d8d4eb902b9e7e7b3c7bf460cfd415e154" Mar 13 14:33:23 crc kubenswrapper[4907]: I0313 14:33:23.506910 4907 scope.go:117] "RemoveContainer" containerID="e3a0771ba1e8115d911c740ed70ee5fca37f11ff18ec41bfab54c46258690b83" Mar 13 14:33:27 crc kubenswrapper[4907]: I0313 14:33:27.782215 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:27 crc kubenswrapper[4907]: E0313 14:33:27.782865 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:39 crc kubenswrapper[4907]: I0313 14:33:39.782423 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:39 crc kubenswrapper[4907]: E0313 14:33:39.783000 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:33:51 crc kubenswrapper[4907]: I0313 14:33:51.787385 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:33:51 crc kubenswrapper[4907]: E0313 14:33:51.788234 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.139806 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:34:00 crc kubenswrapper[4907]: E0313 14:34:00.140720 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-utilities" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140735 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-utilities" Mar 13 14:34:00 crc kubenswrapper[4907]: E0313 14:34:00.140752 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-content" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="extract-content" Mar 13 14:34:00 crc kubenswrapper[4907]: E0313 14:34:00.140786 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140793 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.140997 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf79c546-2d9a-4eba-bf86-f78d196bc557" containerName="registry-server" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.141551 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.143956 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.143970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.143975 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.150111 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.329457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"auto-csr-approver-29556874-chrll\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.430459 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"auto-csr-approver-29556874-chrll\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.450959 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"auto-csr-approver-29556874-chrll\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.477136 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.917631 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:34:00 crc kubenswrapper[4907]: I0313 14:34:00.967153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556874-chrll" event={"ID":"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e","Type":"ContainerStarted","Data":"ed8867f760fa43ecca52e81252b98798addeb34b9728f71dd81a483c15c6d605"} Mar 13 14:34:03 crc kubenswrapper[4907]: I0313 14:34:03.994989 4907 generic.go:334] "Generic (PLEG): container finished" podID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerID="6a420a24c23cba094e67bdf8ae1b0ab87504764d74798dee0bc7e8921659b3b8" exitCode=0 Mar 13 14:34:03 crc kubenswrapper[4907]: I0313 14:34:03.995363 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556874-chrll" event={"ID":"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e","Type":"ContainerDied","Data":"6a420a24c23cba094e67bdf8ae1b0ab87504764d74798dee0bc7e8921659b3b8"} Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.293452 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.398416 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") pod \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\" (UID: \"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e\") " Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.403111 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6" (OuterVolumeSpecName: "kube-api-access-gdsq6") pod "bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" (UID: "bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e"). InnerVolumeSpecName "kube-api-access-gdsq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:34:05 crc kubenswrapper[4907]: I0313 14:34:05.499585 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdsq6\" (UniqueName: \"kubernetes.io/projected/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e-kube-api-access-gdsq6\") on node \"crc\" DevicePath \"\"" Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.007953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556874-chrll" event={"ID":"bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e","Type":"ContainerDied","Data":"ed8867f760fa43ecca52e81252b98798addeb34b9728f71dd81a483c15c6d605"} Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.008002 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed8867f760fa43ecca52e81252b98798addeb34b9728f71dd81a483c15c6d605" Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.008019 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556874-chrll" Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.351793 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.357363 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556868-pzml6"] Mar 13 14:34:06 crc kubenswrapper[4907]: I0313 14:34:06.782300 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:06 crc kubenswrapper[4907]: E0313 14:34:06.782990 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:07 crc kubenswrapper[4907]: I0313 14:34:07.791440 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6301088c-1715-4d57-a9b8-1ea9f7128560" path="/var/lib/kubelet/pods/6301088c-1715-4d57-a9b8-1ea9f7128560/volumes" Mar 13 14:34:17 crc kubenswrapper[4907]: I0313 14:34:17.782471 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:17 crc kubenswrapper[4907]: E0313 14:34:17.783253 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.760748 4907 scope.go:117] "RemoveContainer" containerID="57e7ca150948c3586cf56019f626e80fe0c0e49f021cce99a8bbf1bcfa081624" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.781016 4907 scope.go:117] "RemoveContainer" containerID="2ef6cd9eb88308a93f1d78fb9954045701e6f121a3c20317549b8db40b280bd4" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.804295 4907 scope.go:117] "RemoveContainer" containerID="3f5c4dcc986bdd72e19fe37fafad59be240102706caf33d75e722b2f347bd3e3" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.837099 4907 scope.go:117] "RemoveContainer" containerID="bb7f9aa2bb828fff3a17bc00a23c15ddb0d4b9b7ae4d6e61a19fc6386d089dbb" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.853817 4907 scope.go:117] "RemoveContainer" containerID="6ad9081a27708afd67b77cedc2aa19d22a6afb40f65f24b3a9da7ca3a519bf38" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.869093 4907 scope.go:117] "RemoveContainer" containerID="5feda018c451c4b446f95e35e5721969c1c913a7887bdf916e5ea2193df476ec" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.909344 4907 scope.go:117] "RemoveContainer" containerID="2cbce92733c8c36aab8d17f31bd5085e394cb7fec21de278bf9301bd315d6bae" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.926025 4907 scope.go:117] "RemoveContainer" containerID="52f2356a35bf1ec5bd881f141be785ea7a9d11be6b6e8da92a08cbad79935d8d" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.979008 4907 scope.go:117] "RemoveContainer" containerID="9fe21d916b51d235fb7ba28d7a6934f6f8ffaa95ccce23ab15efee2b6e69acd5" Mar 13 14:34:23 crc kubenswrapper[4907]: I0313 14:34:23.997081 4907 scope.go:117] "RemoveContainer" containerID="a07507237e2303a227977bfc1acdae544607145ad994269ea3936668b9e9c650" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.011121 4907 scope.go:117] "RemoveContainer" containerID="598b6fbc5376df701a762ba6746b37137dc421681ebf472e4298098628662035" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.026644 4907 scope.go:117] "RemoveContainer" containerID="8ce806c1853f002c1ca5aee8d1ea69c43b2c38c4c666d7df49a773673406c573" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.068673 4907 scope.go:117] "RemoveContainer" containerID="107f7252eaea9f43079c72f27a3a08aa9bdfea92d244d423f8695cb45030deb6" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.084365 4907 scope.go:117] "RemoveContainer" containerID="ed14d79ba88da2a61d40c8ed633b8377f32a153aac0ee43765b80eb04fb89188" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.103684 4907 scope.go:117] "RemoveContainer" containerID="13e164891acf22f2bb718fb70e94321cdc023003038bdb8dbc28a6f2b84707b6" Mar 13 14:34:24 crc kubenswrapper[4907]: I0313 14:34:24.122629 4907 scope.go:117] "RemoveContainer" containerID="e156e90b4d1cbea6f3a55f14df88e09d8aeafa1d0449cf8de460be879c418cf0" Mar 13 14:34:28 crc kubenswrapper[4907]: I0313 14:34:28.783542 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:28 crc kubenswrapper[4907]: E0313 14:34:28.784494 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:40 crc kubenswrapper[4907]: I0313 14:34:40.782646 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:40 crc kubenswrapper[4907]: E0313 14:34:40.784334 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:34:55 crc kubenswrapper[4907]: I0313 14:34:55.783336 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:34:55 crc kubenswrapper[4907]: E0313 14:34:55.784936 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:10 crc kubenswrapper[4907]: I0313 14:35:10.783220 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:10 crc kubenswrapper[4907]: E0313 14:35:10.784430 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:21 crc kubenswrapper[4907]: I0313 14:35:21.782649 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:21 crc kubenswrapper[4907]: E0313 14:35:21.784381 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.376614 4907 scope.go:117] "RemoveContainer" containerID="7dd9875251f25eba8556dcd6141b77fd52b2025f417f75c1f8e455693aab178b" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.397176 4907 scope.go:117] "RemoveContainer" containerID="dd9f647265d73fb65018c8a51bfa4b5a5aa86cde573e29956aaf3479042709d4" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.441004 4907 scope.go:117] "RemoveContainer" containerID="d935db5c32cf1af5583418c37d51248a6d17b0514af1c3fa70469ca109747682" Mar 13 14:35:24 crc kubenswrapper[4907]: I0313 14:35:24.477063 4907 scope.go:117] "RemoveContainer" containerID="28a2d9fc7d22dcae7fbbc67bc1cf31d4d10a281e4e795b4b2a25e7a449ecff2c" Mar 13 14:35:34 crc kubenswrapper[4907]: I0313 14:35:34.783534 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:34 crc kubenswrapper[4907]: E0313 14:35:34.784770 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:35:49 crc kubenswrapper[4907]: I0313 14:35:49.782667 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:35:49 crc kubenswrapper[4907]: E0313 14:35:49.784784 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.150533 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:36:00 crc kubenswrapper[4907]: E0313 14:36:00.151505 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerName="oc" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.151523 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerName="oc" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.151698 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" containerName="oc" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.152328 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.156119 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.156492 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.157705 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.161481 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.216567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"auto-csr-approver-29556876-dwnhg\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.318189 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"auto-csr-approver-29556876-dwnhg\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.345110 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"auto-csr-approver-29556876-dwnhg\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.469525 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:00 crc kubenswrapper[4907]: I0313 14:36:00.943867 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:36:01 crc kubenswrapper[4907]: I0313 14:36:01.786607 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:01 crc kubenswrapper[4907]: E0313 14:36:01.787195 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:01 crc kubenswrapper[4907]: I0313 14:36:01.953568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" event={"ID":"d32570e6-4671-4b95-b312-f4af6c4a1478","Type":"ContainerStarted","Data":"46040338384cd3899901e9d87c65468c58bf61984fc713bee92d18141877dbab"} Mar 13 14:36:02 crc kubenswrapper[4907]: I0313 14:36:02.962702 4907 generic.go:334] "Generic (PLEG): container finished" podID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerID="ec22972508d423bdab0d6a15e5b2be7962b2967f440897919fe9018a840dad0f" exitCode=0 Mar 13 14:36:02 crc kubenswrapper[4907]: I0313 14:36:02.963027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" event={"ID":"d32570e6-4671-4b95-b312-f4af6c4a1478","Type":"ContainerDied","Data":"ec22972508d423bdab0d6a15e5b2be7962b2967f440897919fe9018a840dad0f"} Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.399821 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.573527 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") pod \"d32570e6-4671-4b95-b312-f4af6c4a1478\" (UID: \"d32570e6-4671-4b95-b312-f4af6c4a1478\") " Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.578943 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49" (OuterVolumeSpecName: "kube-api-access-65b49") pod "d32570e6-4671-4b95-b312-f4af6c4a1478" (UID: "d32570e6-4671-4b95-b312-f4af6c4a1478"). InnerVolumeSpecName "kube-api-access-65b49". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.675173 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65b49\" (UniqueName: \"kubernetes.io/projected/d32570e6-4671-4b95-b312-f4af6c4a1478-kube-api-access-65b49\") on node \"crc\" DevicePath \"\"" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.986928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" event={"ID":"d32570e6-4671-4b95-b312-f4af6c4a1478","Type":"ContainerDied","Data":"46040338384cd3899901e9d87c65468c58bf61984fc713bee92d18141877dbab"} Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.987000 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46040338384cd3899901e9d87c65468c58bf61984fc713bee92d18141877dbab" Mar 13 14:36:04 crc kubenswrapper[4907]: I0313 14:36:04.986953 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556876-dwnhg" Mar 13 14:36:05 crc kubenswrapper[4907]: I0313 14:36:05.461419 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:36:05 crc kubenswrapper[4907]: I0313 14:36:05.466949 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556870-8mlkz"] Mar 13 14:36:05 crc kubenswrapper[4907]: I0313 14:36:05.793568 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7993f31-1072-4c83-a3ac-16dca5eb04a3" path="/var/lib/kubelet/pods/e7993f31-1072-4c83-a3ac-16dca5eb04a3/volumes" Mar 13 14:36:16 crc kubenswrapper[4907]: I0313 14:36:16.783478 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:16 crc kubenswrapper[4907]: E0313 14:36:16.784823 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:24 crc kubenswrapper[4907]: I0313 14:36:24.558450 4907 scope.go:117] "RemoveContainer" containerID="714551bd9afe88fbb37b8d1ba1d86847559c10eb4afc2d8c8cade9bb4be375ee" Mar 13 14:36:24 crc kubenswrapper[4907]: I0313 14:36:24.612190 4907 scope.go:117] "RemoveContainer" containerID="93b983c69ed0cdc38b2c77a7998230bc3acd8a611b28bc2ce28fc6ae6eacb20d" Mar 13 14:36:29 crc kubenswrapper[4907]: I0313 14:36:29.782465 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:29 crc kubenswrapper[4907]: E0313 14:36:29.783317 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:43 crc kubenswrapper[4907]: I0313 14:36:43.782839 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:43 crc kubenswrapper[4907]: E0313 14:36:43.783579 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:36:56 crc kubenswrapper[4907]: I0313 14:36:56.782532 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:36:56 crc kubenswrapper[4907]: E0313 14:36:56.783491 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:37:10 crc kubenswrapper[4907]: I0313 14:37:10.782439 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:37:10 crc kubenswrapper[4907]: E0313 14:37:10.783294 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:37:25 crc kubenswrapper[4907]: I0313 14:37:25.783258 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:37:26 crc kubenswrapper[4907]: I0313 14:37:26.676273 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd"} Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.150870 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:38:00 crc kubenswrapper[4907]: E0313 14:38:00.151941 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerName="oc" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.151964 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerName="oc" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.152201 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" containerName="oc" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.152764 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.157471 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.157676 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.158201 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.163001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.206799 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"auto-csr-approver-29556878-fqsf6\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.307672 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"auto-csr-approver-29556878-fqsf6\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.327282 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"auto-csr-approver-29556878-fqsf6\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.478100 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.763652 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.765951 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:38:00 crc kubenswrapper[4907]: I0313 14:38:00.943167 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" event={"ID":"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3","Type":"ContainerStarted","Data":"efa9657ca20cfe1e7c6d4d50a0574daaf6c8bf76cb2e8a66510138fcb99666c3"} Mar 13 14:38:02 crc kubenswrapper[4907]: I0313 14:38:02.963779 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerID="3d950127b4213b5d0d64f24ed2cfdf8b4f35b40aa5e7072728ac45679d7fc695" exitCode=0 Mar 13 14:38:02 crc kubenswrapper[4907]: I0313 14:38:02.963838 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" event={"ID":"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3","Type":"ContainerDied","Data":"3d950127b4213b5d0d64f24ed2cfdf8b4f35b40aa5e7072728ac45679d7fc695"} Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.288496 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.369705 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") pod \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\" (UID: \"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3\") " Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.378265 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c" (OuterVolumeSpecName: "kube-api-access-nkf9c") pod "ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" (UID: "ec5b140c-a271-4a8e-a9af-8e7c5bddcad3"). InnerVolumeSpecName "kube-api-access-nkf9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.471433 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkf9c\" (UniqueName: \"kubernetes.io/projected/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3-kube-api-access-nkf9c\") on node \"crc\" DevicePath \"\"" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.980944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" event={"ID":"ec5b140c-a271-4a8e-a9af-8e7c5bddcad3","Type":"ContainerDied","Data":"efa9657ca20cfe1e7c6d4d50a0574daaf6c8bf76cb2e8a66510138fcb99666c3"} Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.981123 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efa9657ca20cfe1e7c6d4d50a0574daaf6c8bf76cb2e8a66510138fcb99666c3" Mar 13 14:38:04 crc kubenswrapper[4907]: I0313 14:38:04.981179 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556878-fqsf6" Mar 13 14:38:05 crc kubenswrapper[4907]: I0313 14:38:05.345971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:38:05 crc kubenswrapper[4907]: I0313 14:38:05.351766 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556872-srgh9"] Mar 13 14:38:05 crc kubenswrapper[4907]: I0313 14:38:05.799002 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32297c61-6461-4357-832f-9de703d88000" path="/var/lib/kubelet/pods/32297c61-6461-4357-832f-9de703d88000/volumes" Mar 13 14:38:24 crc kubenswrapper[4907]: I0313 14:38:24.725010 4907 scope.go:117] "RemoveContainer" containerID="8eea35f2339f97a9cd81df041bdf950f47b0d8ec3a69dea570886bdde6ad88a7" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.042016 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.042652 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.432216 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:39:48 crc kubenswrapper[4907]: E0313 14:39:48.433088 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerName="oc" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.433329 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerName="oc" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.433867 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" containerName="oc" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.436327 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.442009 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.572215 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.572344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.572394 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.674244 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.674757 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.675055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.675174 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.675820 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.713137 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"redhat-marketplace-mqhct\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:48 crc kubenswrapper[4907]: I0313 14:39:48.782120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.196744 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.922780 4907 generic.go:334] "Generic (PLEG): container finished" podID="631842f8-7a11-4ec0-bab3-5793a949806d" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" exitCode=0 Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.922841 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f"} Mar 13 14:39:49 crc kubenswrapper[4907]: I0313 14:39:49.922950 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerStarted","Data":"3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8"} Mar 13 14:39:50 crc kubenswrapper[4907]: I0313 14:39:50.932853 4907 generic.go:334] "Generic (PLEG): container finished" podID="631842f8-7a11-4ec0-bab3-5793a949806d" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" exitCode=0 Mar 13 14:39:50 crc kubenswrapper[4907]: I0313 14:39:50.932915 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759"} Mar 13 14:39:52 crc kubenswrapper[4907]: I0313 14:39:52.953253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerStarted","Data":"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2"} Mar 13 14:39:52 crc kubenswrapper[4907]: I0313 14:39:52.979294 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mqhct" podStartSLOduration=2.545098287 podStartE2EDuration="4.97927148s" podCreationTimestamp="2026-03-13 14:39:48 +0000 UTC" firstStartedPulling="2026-03-13 14:39:49.925841359 +0000 UTC m=+2088.825629088" lastFinishedPulling="2026-03-13 14:39:52.360014562 +0000 UTC m=+2091.259802281" observedRunningTime="2026-03-13 14:39:52.976401087 +0000 UTC m=+2091.876188816" watchObservedRunningTime="2026-03-13 14:39:52.97927148 +0000 UTC m=+2091.879059179" Mar 13 14:39:58 crc kubenswrapper[4907]: I0313 14:39:58.782313 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:58 crc kubenswrapper[4907]: I0313 14:39:58.782719 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:58 crc kubenswrapper[4907]: I0313 14:39:58.838397 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:59 crc kubenswrapper[4907]: I0313 14:39:59.060928 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:39:59 crc kubenswrapper[4907]: I0313 14:39:59.116742 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.148674 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.149864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.155835 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.156124 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.156235 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.185781 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.261647 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"auto-csr-approver-29556880-wxdzr\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.362765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"auto-csr-approver-29556880-wxdzr\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.387777 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"auto-csr-approver-29556880-wxdzr\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.481079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:00 crc kubenswrapper[4907]: I0313 14:40:00.956459 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:40:00 crc kubenswrapper[4907]: W0313 14:40:00.967528 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20dce4c7_7cbf_4e49_bd0e_93ee53d72551.slice/crio-879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80 WatchSource:0}: Error finding container 879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80: Status 404 returned error can't find the container with id 879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80 Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.019479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" event={"ID":"20dce4c7-7cbf-4e49-bd0e-93ee53d72551","Type":"ContainerStarted","Data":"879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80"} Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.019604 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mqhct" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" containerID="cri-o://acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" gracePeriod=2 Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.462560 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.586836 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") pod \"631842f8-7a11-4ec0-bab3-5793a949806d\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.587089 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") pod \"631842f8-7a11-4ec0-bab3-5793a949806d\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.587290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") pod \"631842f8-7a11-4ec0-bab3-5793a949806d\" (UID: \"631842f8-7a11-4ec0-bab3-5793a949806d\") " Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.588655 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities" (OuterVolumeSpecName: "utilities") pod "631842f8-7a11-4ec0-bab3-5793a949806d" (UID: "631842f8-7a11-4ec0-bab3-5793a949806d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.594649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9" (OuterVolumeSpecName: "kube-api-access-9ffj9") pod "631842f8-7a11-4ec0-bab3-5793a949806d" (UID: "631842f8-7a11-4ec0-bab3-5793a949806d"). InnerVolumeSpecName "kube-api-access-9ffj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.625458 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "631842f8-7a11-4ec0-bab3-5793a949806d" (UID: "631842f8-7a11-4ec0-bab3-5793a949806d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.689608 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ffj9\" (UniqueName: \"kubernetes.io/projected/631842f8-7a11-4ec0-bab3-5793a949806d-kube-api-access-9ffj9\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.689670 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:01 crc kubenswrapper[4907]: I0313 14:40:01.689690 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/631842f8-7a11-4ec0-bab3-5793a949806d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029051 4907 generic.go:334] "Generic (PLEG): container finished" podID="631842f8-7a11-4ec0-bab3-5793a949806d" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" exitCode=0 Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029109 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqhct" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2"} Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029465 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqhct" event={"ID":"631842f8-7a11-4ec0-bab3-5793a949806d","Type":"ContainerDied","Data":"3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8"} Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.029488 4907 scope.go:117] "RemoveContainer" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.052785 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.058730 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqhct"] Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.066092 4907 scope.go:117] "RemoveContainer" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.083670 4907 scope.go:117] "RemoveContainer" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.110689 4907 scope.go:117] "RemoveContainer" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" Mar 13 14:40:02 crc kubenswrapper[4907]: E0313 14:40:02.111259 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2\": container with ID starting with acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2 not found: ID does not exist" containerID="acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111359 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2"} err="failed to get container status \"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2\": rpc error: code = NotFound desc = could not find container \"acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2\": container with ID starting with acb00c615a16c4afc4f4512ed580341d6c225c7726089602e71b32f317ded8e2 not found: ID does not exist" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111439 4907 scope.go:117] "RemoveContainer" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" Mar 13 14:40:02 crc kubenswrapper[4907]: E0313 14:40:02.111944 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759\": container with ID starting with fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759 not found: ID does not exist" containerID="fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111972 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759"} err="failed to get container status \"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759\": rpc error: code = NotFound desc = could not find container \"fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759\": container with ID starting with fe74f9a5411a632ebac06894970db3a7519b223f6ba533c92f921f1566e8a759 not found: ID does not exist" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.111990 4907 scope.go:117] "RemoveContainer" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" Mar 13 14:40:02 crc kubenswrapper[4907]: E0313 14:40:02.112349 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f\": container with ID starting with 8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f not found: ID does not exist" containerID="8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f" Mar 13 14:40:02 crc kubenswrapper[4907]: I0313 14:40:02.112374 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f"} err="failed to get container status \"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f\": rpc error: code = NotFound desc = could not find container \"8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f\": container with ID starting with 8b6db5a70191fecc20fe10365c72f6b6153346de6e602aa9933891080440586f not found: ID does not exist" Mar 13 14:40:03 crc kubenswrapper[4907]: I0313 14:40:03.043786 4907 generic.go:334] "Generic (PLEG): container finished" podID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerID="fc46a881237631d2b078045b48ba1243ef9d84385c8036fc8ea4400ece996277" exitCode=0 Mar 13 14:40:03 crc kubenswrapper[4907]: I0313 14:40:03.043946 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" event={"ID":"20dce4c7-7cbf-4e49-bd0e-93ee53d72551","Type":"ContainerDied","Data":"fc46a881237631d2b078045b48ba1243ef9d84385c8036fc8ea4400ece996277"} Mar 13 14:40:03 crc kubenswrapper[4907]: I0313 14:40:03.807851 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" path="/var/lib/kubelet/pods/631842f8-7a11-4ec0-bab3-5793a949806d/volumes" Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.349327 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.434141 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") pod \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\" (UID: \"20dce4c7-7cbf-4e49-bd0e-93ee53d72551\") " Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.442543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk" (OuterVolumeSpecName: "kube-api-access-nj4vk") pod "20dce4c7-7cbf-4e49-bd0e-93ee53d72551" (UID: "20dce4c7-7cbf-4e49-bd0e-93ee53d72551"). InnerVolumeSpecName "kube-api-access-nj4vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:40:04 crc kubenswrapper[4907]: I0313 14:40:04.536116 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj4vk\" (UniqueName: \"kubernetes.io/projected/20dce4c7-7cbf-4e49-bd0e-93ee53d72551-kube-api-access-nj4vk\") on node \"crc\" DevicePath \"\"" Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.065038 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" event={"ID":"20dce4c7-7cbf-4e49-bd0e-93ee53d72551","Type":"ContainerDied","Data":"879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80"} Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.065085 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556880-wxdzr" Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.065092 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="879e5062915a12a2855140a77d7a884e87419cf083062608de3e6cfa55590d80" Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.415336 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.421123 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556874-chrll"] Mar 13 14:40:05 crc kubenswrapper[4907]: I0313 14:40:05.792980 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e" path="/var/lib/kubelet/pods/bb80c3ca-ddb4-49fe-8b24-5dc43269ab1e/volumes" Mar 13 14:40:08 crc kubenswrapper[4907]: E0313 14:40:08.225242 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:18 crc kubenswrapper[4907]: I0313 14:40:18.041961 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:40:18 crc kubenswrapper[4907]: I0313 14:40:18.042710 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:40:18 crc kubenswrapper[4907]: E0313 14:40:18.432648 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:24 crc kubenswrapper[4907]: I0313 14:40:24.805206 4907 scope.go:117] "RemoveContainer" containerID="6a420a24c23cba094e67bdf8ae1b0ab87504764d74798dee0bc7e8921659b3b8" Mar 13 14:40:28 crc kubenswrapper[4907]: E0313 14:40:28.626060 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:38 crc kubenswrapper[4907]: E0313 14:40:38.805560 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.042219 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.042637 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.042678 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.043108 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.043149 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd" gracePeriod=600 Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421640 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd" exitCode=0 Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd"} Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421921 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b"} Mar 13 14:40:48 crc kubenswrapper[4907]: I0313 14:40:48.421958 4907 scope.go:117] "RemoveContainer" containerID="a09a3bec467b7e275495edb5827150af756bcc256911e7e06cbae5765922b5e2" Mar 13 14:40:49 crc kubenswrapper[4907]: E0313 14:40:49.011314 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache]" Mar 13 14:40:59 crc kubenswrapper[4907]: E0313 14:40:59.191465 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice/crio-3849a706a8af92bea878d778fade40002b62f87b6f1a978975072a05bbe7b2c8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631842f8_7a11_4ec0_bab3_5793a949806d.slice\": RecentStats: unable to find data in memory cache]" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.494127 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495515 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerName="oc" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495543 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerName="oc" Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495565 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-utilities" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-utilities" Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495595 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-content" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495604 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="extract-content" Mar 13 14:41:31 crc kubenswrapper[4907]: E0313 14:41:31.495636 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495646 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495831 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" containerName="oc" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.495901 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="631842f8-7a11-4ec0-bab3-5793a949806d" containerName="registry-server" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.498571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.503932 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.511322 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.511381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.511416 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613119 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613188 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613454 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.613800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.636349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"redhat-operators-gg9hb\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:31 crc kubenswrapper[4907]: I0313 14:41:31.816952 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.276645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.757840 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" exitCode=0 Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.758036 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1"} Mar 13 14:41:32 crc kubenswrapper[4907]: I0313 14:41:32.759196 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerStarted","Data":"7e29cff77ed5e8b8c0c9fb8223d475e0bf231b38d49b85e0d18a0eff7bad0daf"} Mar 13 14:41:33 crc kubenswrapper[4907]: I0313 14:41:33.767139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerStarted","Data":"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422"} Mar 13 14:41:34 crc kubenswrapper[4907]: I0313 14:41:34.780074 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" exitCode=0 Mar 13 14:41:34 crc kubenswrapper[4907]: I0313 14:41:34.780127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422"} Mar 13 14:41:35 crc kubenswrapper[4907]: I0313 14:41:35.790182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerStarted","Data":"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269"} Mar 13 14:41:35 crc kubenswrapper[4907]: I0313 14:41:35.809133 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gg9hb" podStartSLOduration=2.364032716 podStartE2EDuration="4.809114429s" podCreationTimestamp="2026-03-13 14:41:31 +0000 UTC" firstStartedPulling="2026-03-13 14:41:32.760064205 +0000 UTC m=+2191.659851894" lastFinishedPulling="2026-03-13 14:41:35.205145918 +0000 UTC m=+2194.104933607" observedRunningTime="2026-03-13 14:41:35.805193203 +0000 UTC m=+2194.704980902" watchObservedRunningTime="2026-03-13 14:41:35.809114429 +0000 UTC m=+2194.708902128" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.817559 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.819043 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.867142 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:41 crc kubenswrapper[4907]: I0313 14:41:41.910426 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:42 crc kubenswrapper[4907]: I0313 14:41:42.099508 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:43 crc kubenswrapper[4907]: I0313 14:41:43.856188 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gg9hb" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" containerID="cri-o://1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" gracePeriod=2 Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.220640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.390861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") pod \"7ff81c42-6310-47a8-a691-5a7193a1f806\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.390987 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") pod \"7ff81c42-6310-47a8-a691-5a7193a1f806\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.392098 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") pod \"7ff81c42-6310-47a8-a691-5a7193a1f806\" (UID: \"7ff81c42-6310-47a8-a691-5a7193a1f806\") " Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.392810 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities" (OuterVolumeSpecName: "utilities") pod "7ff81c42-6310-47a8-a691-5a7193a1f806" (UID: "7ff81c42-6310-47a8-a691-5a7193a1f806"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.401547 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9" (OuterVolumeSpecName: "kube-api-access-2c5n9") pod "7ff81c42-6310-47a8-a691-5a7193a1f806" (UID: "7ff81c42-6310-47a8-a691-5a7193a1f806"). InnerVolumeSpecName "kube-api-access-2c5n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.494004 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2c5n9\" (UniqueName: \"kubernetes.io/projected/7ff81c42-6310-47a8-a691-5a7193a1f806-kube-api-access-2c5n9\") on node \"crc\" DevicePath \"\"" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.494040 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865713 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" exitCode=0 Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269"} Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865787 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gg9hb" event={"ID":"7ff81c42-6310-47a8-a691-5a7193a1f806","Type":"ContainerDied","Data":"7e29cff77ed5e8b8c0c9fb8223d475e0bf231b38d49b85e0d18a0eff7bad0daf"} Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.865806 4907 scope.go:117] "RemoveContainer" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.867013 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gg9hb" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.884661 4907 scope.go:117] "RemoveContainer" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.899871 4907 scope.go:117] "RemoveContainer" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.922303 4907 scope.go:117] "RemoveContainer" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" Mar 13 14:41:44 crc kubenswrapper[4907]: E0313 14:41:44.922754 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269\": container with ID starting with 1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269 not found: ID does not exist" containerID="1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.922814 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269"} err="failed to get container status \"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269\": rpc error: code = NotFound desc = could not find container \"1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269\": container with ID starting with 1554edf4b3f861c840f593511dedd2375f020bd21204d76082b3b72355050269 not found: ID does not exist" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.922855 4907 scope.go:117] "RemoveContainer" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" Mar 13 14:41:44 crc kubenswrapper[4907]: E0313 14:41:44.923284 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422\": container with ID starting with b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422 not found: ID does not exist" containerID="b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.923319 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422"} err="failed to get container status \"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422\": rpc error: code = NotFound desc = could not find container \"b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422\": container with ID starting with b5a0413a722d5577be71421092aa45b7b0b6b423078bf9472fdcb79739eef422 not found: ID does not exist" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.923341 4907 scope.go:117] "RemoveContainer" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" Mar 13 14:41:44 crc kubenswrapper[4907]: E0313 14:41:44.923630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1\": container with ID starting with 04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1 not found: ID does not exist" containerID="04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1" Mar 13 14:41:44 crc kubenswrapper[4907]: I0313 14:41:44.923666 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1"} err="failed to get container status \"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1\": rpc error: code = NotFound desc = could not find container \"04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1\": container with ID starting with 04c538f4410703a67c9a1f6f70fb480854f64cb603b98dc4fb30a4da10e1cfe1 not found: ID does not exist" Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.556679 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ff81c42-6310-47a8-a691-5a7193a1f806" (UID: "7ff81c42-6310-47a8-a691-5a7193a1f806"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.610996 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ff81c42-6310-47a8-a691-5a7193a1f806-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.804243 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:45 crc kubenswrapper[4907]: I0313 14:41:45.811168 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gg9hb"] Mar 13 14:41:47 crc kubenswrapper[4907]: I0313 14:41:47.792617 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" path="/var/lib/kubelet/pods/7ff81c42-6310-47a8-a691-5a7193a1f806/volumes" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.318970 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:41:49 crc kubenswrapper[4907]: E0313 14:41:49.319947 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-content" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320251 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-content" Mar 13 14:41:49 crc kubenswrapper[4907]: E0313 14:41:49.320342 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-utilities" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320417 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="extract-utilities" Mar 13 14:41:49 crc kubenswrapper[4907]: E0313 14:41:49.320494 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320576 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.320800 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ff81c42-6310-47a8-a691-5a7193a1f806" containerName="registry-server" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.322360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.336633 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.461609 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.461655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.461675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.562838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.562932 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.562975 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.563397 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.563597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.618029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"certified-operators-kh8q2\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:49 crc kubenswrapper[4907]: I0313 14:41:49.642004 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.150076 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.905586 4907 generic.go:334] "Generic (PLEG): container finished" podID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" exitCode=0 Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.905661 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44"} Mar 13 14:41:50 crc kubenswrapper[4907]: I0313 14:41:50.905932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerStarted","Data":"410b8cd335bd7034fb5a1e0759b6ec51be56bf7900b40557e8a393482878522f"} Mar 13 14:41:51 crc kubenswrapper[4907]: I0313 14:41:51.914307 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerStarted","Data":"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31"} Mar 13 14:41:52 crc kubenswrapper[4907]: I0313 14:41:52.921807 4907 generic.go:334] "Generic (PLEG): container finished" podID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" exitCode=0 Mar 13 14:41:52 crc kubenswrapper[4907]: I0313 14:41:52.921863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31"} Mar 13 14:41:53 crc kubenswrapper[4907]: I0313 14:41:53.934936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerStarted","Data":"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64"} Mar 13 14:41:53 crc kubenswrapper[4907]: I0313 14:41:53.961325 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kh8q2" podStartSLOduration=2.124273589 podStartE2EDuration="4.961295885s" podCreationTimestamp="2026-03-13 14:41:49 +0000 UTC" firstStartedPulling="2026-03-13 14:41:50.907080918 +0000 UTC m=+2209.806868607" lastFinishedPulling="2026-03-13 14:41:53.744103214 +0000 UTC m=+2212.643890903" observedRunningTime="2026-03-13 14:41:53.960161253 +0000 UTC m=+2212.859948942" watchObservedRunningTime="2026-03-13 14:41:53.961295885 +0000 UTC m=+2212.861083614" Mar 13 14:41:59 crc kubenswrapper[4907]: I0313 14:41:59.649119 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:59 crc kubenswrapper[4907]: I0313 14:41:59.649805 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:41:59 crc kubenswrapper[4907]: I0313 14:41:59.692127 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.020515 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.065903 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.143402 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.144258 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.146932 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.147426 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.147433 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.158459 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.212997 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"auto-csr-approver-29556882-c7dc8\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.315795 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"auto-csr-approver-29556882-c7dc8\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.336935 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"auto-csr-approver-29556882-c7dc8\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.467025 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.897340 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:42:00 crc kubenswrapper[4907]: W0313 14:42:00.904952 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81733636_fc5e_4687_9ceb_0dd2e18fa15a.slice/crio-47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76 WatchSource:0}: Error finding container 47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76: Status 404 returned error can't find the container with id 47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76 Mar 13 14:42:00 crc kubenswrapper[4907]: I0313 14:42:00.985594 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" event={"ID":"81733636-fc5e-4687-9ceb-0dd2e18fa15a","Type":"ContainerStarted","Data":"47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76"} Mar 13 14:42:01 crc kubenswrapper[4907]: I0313 14:42:01.994427 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kh8q2" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" containerID="cri-o://838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" gracePeriod=2 Mar 13 14:42:02 crc kubenswrapper[4907]: I0313 14:42:02.926270 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.006358 4907 generic.go:334] "Generic (PLEG): container finished" podID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerID="888132585d34376894f6b13900aa0293c6191b881a23e9abf82e3ca297b39ec9" exitCode=0 Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.006411 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" event={"ID":"81733636-fc5e-4687-9ceb-0dd2e18fa15a","Type":"ContainerDied","Data":"888132585d34376894f6b13900aa0293c6191b881a23e9abf82e3ca297b39ec9"} Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009730 4907 generic.go:334] "Generic (PLEG): container finished" podID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" exitCode=0 Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009770 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64"} Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009792 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kh8q2" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009830 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kh8q2" event={"ID":"bbd9d05a-e055-4974-b018-f1d8fdd10178","Type":"ContainerDied","Data":"410b8cd335bd7034fb5a1e0759b6ec51be56bf7900b40557e8a393482878522f"} Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.009861 4907 scope.go:117] "RemoveContainer" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.034816 4907 scope.go:117] "RemoveContainer" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.050324 4907 scope.go:117] "RemoveContainer" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.062492 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") pod \"bbd9d05a-e055-4974-b018-f1d8fdd10178\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.062577 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") pod \"bbd9d05a-e055-4974-b018-f1d8fdd10178\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.062663 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") pod \"bbd9d05a-e055-4974-b018-f1d8fdd10178\" (UID: \"bbd9d05a-e055-4974-b018-f1d8fdd10178\") " Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.063366 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities" (OuterVolumeSpecName: "utilities") pod "bbd9d05a-e055-4974-b018-f1d8fdd10178" (UID: "bbd9d05a-e055-4974-b018-f1d8fdd10178"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.068944 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf" (OuterVolumeSpecName: "kube-api-access-rmldf") pod "bbd9d05a-e055-4974-b018-f1d8fdd10178" (UID: "bbd9d05a-e055-4974-b018-f1d8fdd10178"). InnerVolumeSpecName "kube-api-access-rmldf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.076985 4907 scope.go:117] "RemoveContainer" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" Mar 13 14:42:03 crc kubenswrapper[4907]: E0313 14:42:03.077463 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64\": container with ID starting with 838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64 not found: ID does not exist" containerID="838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.077500 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64"} err="failed to get container status \"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64\": rpc error: code = NotFound desc = could not find container \"838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64\": container with ID starting with 838f1e0eb2d0fa893d5a1cfcea314424ab3346a5d58f11c35c05c50224a6bf64 not found: ID does not exist" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.077542 4907 scope.go:117] "RemoveContainer" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" Mar 13 14:42:03 crc kubenswrapper[4907]: E0313 14:42:03.078105 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31\": container with ID starting with da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31 not found: ID does not exist" containerID="da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.078129 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31"} err="failed to get container status \"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31\": rpc error: code = NotFound desc = could not find container \"da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31\": container with ID starting with da85c297a9d3589a9b32a44fd9795a8682a003114947247f16b543e60836dd31 not found: ID does not exist" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.078162 4907 scope.go:117] "RemoveContainer" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" Mar 13 14:42:03 crc kubenswrapper[4907]: E0313 14:42:03.078418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44\": container with ID starting with c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44 not found: ID does not exist" containerID="c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.078458 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44"} err="failed to get container status \"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44\": rpc error: code = NotFound desc = could not find container \"c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44\": container with ID starting with c29f21e67b78bfa6321f01394b48fae88a9c4f77cce83e14401fce3ca805ce44 not found: ID does not exist" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.130341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbd9d05a-e055-4974-b018-f1d8fdd10178" (UID: "bbd9d05a-e055-4974-b018-f1d8fdd10178"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.164581 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmldf\" (UniqueName: \"kubernetes.io/projected/bbd9d05a-e055-4974-b018-f1d8fdd10178-kube-api-access-rmldf\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.164609 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.164621 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbd9d05a-e055-4974-b018-f1d8fdd10178-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.361396 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.375308 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kh8q2"] Mar 13 14:42:03 crc kubenswrapper[4907]: I0313 14:42:03.790008 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" path="/var/lib/kubelet/pods/bbd9d05a-e055-4974-b018-f1d8fdd10178/volumes" Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.345776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.486099 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") pod \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\" (UID: \"81733636-fc5e-4687-9ceb-0dd2e18fa15a\") " Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.493493 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt" (OuterVolumeSpecName: "kube-api-access-xkkqt") pod "81733636-fc5e-4687-9ceb-0dd2e18fa15a" (UID: "81733636-fc5e-4687-9ceb-0dd2e18fa15a"). InnerVolumeSpecName "kube-api-access-xkkqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:42:04 crc kubenswrapper[4907]: I0313 14:42:04.588638 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkkqt\" (UniqueName: \"kubernetes.io/projected/81733636-fc5e-4687-9ceb-0dd2e18fa15a-kube-api-access-xkkqt\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.031845 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" event={"ID":"81733636-fc5e-4687-9ceb-0dd2e18fa15a","Type":"ContainerDied","Data":"47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76"} Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.031899 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47339f6bc7f00baa2ec0d1f9fb7f56f3dda07e47348aedc197d070f454b4bb76" Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.031911 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556882-c7dc8" Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.418741 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.424185 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556876-dwnhg"] Mar 13 14:42:05 crc kubenswrapper[4907]: I0313 14:42:05.794667 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d32570e6-4671-4b95-b312-f4af6c4a1478" path="/var/lib/kubelet/pods/d32570e6-4671-4b95-b312-f4af6c4a1478/volumes" Mar 13 14:42:24 crc kubenswrapper[4907]: I0313 14:42:24.923607 4907 scope.go:117] "RemoveContainer" containerID="ec22972508d423bdab0d6a15e5b2be7962b2967f440897919fe9018a840dad0f" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.185426 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186673 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186707 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186724 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-content" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186733 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-content" Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186750 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerName="oc" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerName="oc" Mar 13 14:42:34 crc kubenswrapper[4907]: E0313 14:42:34.186782 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-utilities" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186790 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="extract-utilities" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.186987 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" containerName="oc" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.187018 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbd9d05a-e055-4974-b018-f1d8fdd10178" containerName="registry-server" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.188061 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.201745 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.259812 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxdf7\" (UniqueName: \"kubernetes.io/projected/e3a3b609-3625-4670-b669-d7fd07386be6-kube-api-access-rxdf7\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.259930 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-utilities\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.259965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-catalog-content\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.361345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxdf7\" (UniqueName: \"kubernetes.io/projected/e3a3b609-3625-4670-b669-d7fd07386be6-kube-api-access-rxdf7\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.361455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-utilities\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.361484 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-catalog-content\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.362119 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-catalog-content\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.362244 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3a3b609-3625-4670-b669-d7fd07386be6-utilities\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.384745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxdf7\" (UniqueName: \"kubernetes.io/projected/e3a3b609-3625-4670-b669-d7fd07386be6-kube-api-access-rxdf7\") pod \"community-operators-bwznz\" (UID: \"e3a3b609-3625-4670-b669-d7fd07386be6\") " pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:34 crc kubenswrapper[4907]: I0313 14:42:34.526109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.009946 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.267729 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3a3b609-3625-4670-b669-d7fd07386be6" containerID="d1ed90bc0d96aa6d2be8063d734ff2e5ee2f08e2f5f3cf16acdb090807cbe152" exitCode=0 Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.267798 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerDied","Data":"d1ed90bc0d96aa6d2be8063d734ff2e5ee2f08e2f5f3cf16acdb090807cbe152"} Mar 13 14:42:35 crc kubenswrapper[4907]: I0313 14:42:35.268031 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerStarted","Data":"b12c47624fc9c120d0c9e10ec66f5d5bf123a990a2f6d121f6b346f5dcde3df9"} Mar 13 14:42:39 crc kubenswrapper[4907]: I0313 14:42:39.292992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerStarted","Data":"e9afb226fd8c064f22333c56811e1ee9b967bd4ba350cccbeba6a963b90b37da"} Mar 13 14:42:40 crc kubenswrapper[4907]: I0313 14:42:40.301007 4907 generic.go:334] "Generic (PLEG): container finished" podID="e3a3b609-3625-4670-b669-d7fd07386be6" containerID="e9afb226fd8c064f22333c56811e1ee9b967bd4ba350cccbeba6a963b90b37da" exitCode=0 Mar 13 14:42:40 crc kubenswrapper[4907]: I0313 14:42:40.301047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerDied","Data":"e9afb226fd8c064f22333c56811e1ee9b967bd4ba350cccbeba6a963b90b37da"} Mar 13 14:42:41 crc kubenswrapper[4907]: I0313 14:42:41.309417 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bwznz" event={"ID":"e3a3b609-3625-4670-b669-d7fd07386be6","Type":"ContainerStarted","Data":"0c1015ce501c5ee89732127ca1e7eceae8fce9d0539ede2b5801447664bc10f4"} Mar 13 14:42:41 crc kubenswrapper[4907]: I0313 14:42:41.327842 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bwznz" podStartSLOduration=1.858644712 podStartE2EDuration="7.327826578s" podCreationTimestamp="2026-03-13 14:42:34 +0000 UTC" firstStartedPulling="2026-03-13 14:42:35.269381382 +0000 UTC m=+2254.169169071" lastFinishedPulling="2026-03-13 14:42:40.738563248 +0000 UTC m=+2259.638350937" observedRunningTime="2026-03-13 14:42:41.327536479 +0000 UTC m=+2260.227324178" watchObservedRunningTime="2026-03-13 14:42:41.327826578 +0000 UTC m=+2260.227614267" Mar 13 14:42:44 crc kubenswrapper[4907]: I0313 14:42:44.527232 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:44 crc kubenswrapper[4907]: I0313 14:42:44.527521 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:44 crc kubenswrapper[4907]: I0313 14:42:44.584706 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.387671 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bwznz" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.476956 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bwznz"] Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.504076 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.507982 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bkqkq" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" containerID="cri-o://dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" gracePeriod=2 Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.910760 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.918353 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") pod \"8f55641f-96d1-4abe-b29e-35c03423c86f\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.918422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") pod \"8f55641f-96d1-4abe-b29e-35c03423c86f\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.918601 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") pod \"8f55641f-96d1-4abe-b29e-35c03423c86f\" (UID: \"8f55641f-96d1-4abe-b29e-35c03423c86f\") " Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.920543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities" (OuterVolumeSpecName: "utilities") pod "8f55641f-96d1-4abe-b29e-35c03423c86f" (UID: "8f55641f-96d1-4abe-b29e-35c03423c86f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.931841 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh" (OuterVolumeSpecName: "kube-api-access-blsqh") pod "8f55641f-96d1-4abe-b29e-35c03423c86f" (UID: "8f55641f-96d1-4abe-b29e-35c03423c86f"). InnerVolumeSpecName "kube-api-access-blsqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:42:45 crc kubenswrapper[4907]: I0313 14:42:45.965725 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f55641f-96d1-4abe-b29e-35c03423c86f" (UID: "8f55641f-96d1-4abe-b29e-35c03423c86f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.019796 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blsqh\" (UniqueName: \"kubernetes.io/projected/8f55641f-96d1-4abe-b29e-35c03423c86f-kube-api-access-blsqh\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.019956 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.020021 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f55641f-96d1-4abe-b29e-35c03423c86f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344764 4907 generic.go:334] "Generic (PLEG): container finished" podID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" exitCode=0 Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344836 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bkqkq" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344868 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178"} Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bkqkq" event={"ID":"8f55641f-96d1-4abe-b29e-35c03423c86f","Type":"ContainerDied","Data":"51741992329c7cbf452d56e06ea4b68fc6d3bd2347af300a47210ccc94035b08"} Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.344949 4907 scope.go:117] "RemoveContainer" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.372832 4907 scope.go:117] "RemoveContainer" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.384304 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.389389 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bkqkq"] Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.398252 4907 scope.go:117] "RemoveContainer" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.428739 4907 scope.go:117] "RemoveContainer" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" Mar 13 14:42:46 crc kubenswrapper[4907]: E0313 14:42:46.429097 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178\": container with ID starting with dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178 not found: ID does not exist" containerID="dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429127 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178"} err="failed to get container status \"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178\": rpc error: code = NotFound desc = could not find container \"dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178\": container with ID starting with dd30544066f21dab70ce804b909280851ba1f1bd4d4cb32479861077957a9178 not found: ID does not exist" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429148 4907 scope.go:117] "RemoveContainer" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" Mar 13 14:42:46 crc kubenswrapper[4907]: E0313 14:42:46.429343 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654\": container with ID starting with 7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654 not found: ID does not exist" containerID="7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429363 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654"} err="failed to get container status \"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654\": rpc error: code = NotFound desc = could not find container \"7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654\": container with ID starting with 7fb98f669ca95b8966acee50a69bfe3ca16749a7b12337396bf39c2c672c9654 not found: ID does not exist" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429394 4907 scope.go:117] "RemoveContainer" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" Mar 13 14:42:46 crc kubenswrapper[4907]: E0313 14:42:46.429604 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946\": container with ID starting with 4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946 not found: ID does not exist" containerID="4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946" Mar 13 14:42:46 crc kubenswrapper[4907]: I0313 14:42:46.429625 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946"} err="failed to get container status \"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946\": rpc error: code = NotFound desc = could not find container \"4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946\": container with ID starting with 4a29b590b6d8237b3616d05e54dcdd32085b2501383cd2a01ceaf05da4c45946 not found: ID does not exist" Mar 13 14:42:47 crc kubenswrapper[4907]: I0313 14:42:47.792463 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" path="/var/lib/kubelet/pods/8f55641f-96d1-4abe-b29e-35c03423c86f/volumes" Mar 13 14:42:48 crc kubenswrapper[4907]: I0313 14:42:48.042404 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:42:48 crc kubenswrapper[4907]: I0313 14:42:48.042485 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:43:18 crc kubenswrapper[4907]: I0313 14:43:18.041364 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:43:18 crc kubenswrapper[4907]: I0313 14:43:18.042951 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.042035 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.044054 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.044142 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.045197 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.045288 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" gracePeriod=600 Mar 13 14:43:48 crc kubenswrapper[4907]: E0313 14:43:48.184178 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857141 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" exitCode=0 Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857194 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b"} Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857241 4907 scope.go:117] "RemoveContainer" containerID="3c1970952dcd249074d61677266bef44d3d78378f555ffc20b89b00a033811dd" Mar 13 14:43:48 crc kubenswrapper[4907]: I0313 14:43:48.857782 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:43:48 crc kubenswrapper[4907]: E0313 14:43:48.858076 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.135507 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.138904 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-utilities" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.138986 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-utilities" Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.139014 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-content" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.139037 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="extract-content" Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.139050 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.139059 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.139316 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f55641f-96d1-4abe-b29e-35c03423c86f" containerName="registry-server" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.140507 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.146683 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.146732 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.146761 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.147726 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.243562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"auto-csr-approver-29556884-zzm6t\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.345008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"auto-csr-approver-29556884-zzm6t\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.362170 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"auto-csr-approver-29556884-zzm6t\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.462598 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.738264 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.749714 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.782129 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:00 crc kubenswrapper[4907]: E0313 14:44:00.782352 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:00 crc kubenswrapper[4907]: I0313 14:44:00.963149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" event={"ID":"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc","Type":"ContainerStarted","Data":"2083bf0f478da887209b5b2d38e198b775cc56d66e0058fae942f158d67a7dda"} Mar 13 14:44:02 crc kubenswrapper[4907]: I0313 14:44:02.981476 4907 generic.go:334] "Generic (PLEG): container finished" podID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerID="1d99f9504e5b5c3d1b261df414bc1841776b381d4074946ac4e7f45b3d386557" exitCode=0 Mar 13 14:44:02 crc kubenswrapper[4907]: I0313 14:44:02.981556 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" event={"ID":"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc","Type":"ContainerDied","Data":"1d99f9504e5b5c3d1b261df414bc1841776b381d4074946ac4e7f45b3d386557"} Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.349765 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.512123 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") pod \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\" (UID: \"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc\") " Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.518505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk" (OuterVolumeSpecName: "kube-api-access-b92lk") pod "bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" (UID: "bd8a84e4-33ca-4236-8c28-bf80cf64b6bc"). InnerVolumeSpecName "kube-api-access-b92lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.613958 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b92lk\" (UniqueName: \"kubernetes.io/projected/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc-kube-api-access-b92lk\") on node \"crc\" DevicePath \"\"" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.996951 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" event={"ID":"bd8a84e4-33ca-4236-8c28-bf80cf64b6bc","Type":"ContainerDied","Data":"2083bf0f478da887209b5b2d38e198b775cc56d66e0058fae942f158d67a7dda"} Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.997301 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2083bf0f478da887209b5b2d38e198b775cc56d66e0058fae942f158d67a7dda" Mar 13 14:44:04 crc kubenswrapper[4907]: I0313 14:44:04.997450 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556884-zzm6t" Mar 13 14:44:05 crc kubenswrapper[4907]: I0313 14:44:05.423366 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:44:05 crc kubenswrapper[4907]: I0313 14:44:05.431138 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556878-fqsf6"] Mar 13 14:44:05 crc kubenswrapper[4907]: I0313 14:44:05.796478 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec5b140c-a271-4a8e-a9af-8e7c5bddcad3" path="/var/lib/kubelet/pods/ec5b140c-a271-4a8e-a9af-8e7c5bddcad3/volumes" Mar 13 14:44:15 crc kubenswrapper[4907]: I0313 14:44:15.783134 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:15 crc kubenswrapper[4907]: E0313 14:44:15.784138 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:25 crc kubenswrapper[4907]: I0313 14:44:25.061723 4907 scope.go:117] "RemoveContainer" containerID="3d950127b4213b5d0d64f24ed2cfdf8b4f35b40aa5e7072728ac45679d7fc695" Mar 13 14:44:29 crc kubenswrapper[4907]: I0313 14:44:29.782381 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:29 crc kubenswrapper[4907]: E0313 14:44:29.783118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:43 crc kubenswrapper[4907]: I0313 14:44:43.782717 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:43 crc kubenswrapper[4907]: E0313 14:44:43.783564 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:44:54 crc kubenswrapper[4907]: I0313 14:44:54.783183 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:44:54 crc kubenswrapper[4907]: E0313 14:44:54.784353 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.160436 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 14:45:00 crc kubenswrapper[4907]: E0313 14:45:00.161010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerName="oc" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.161022 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerName="oc" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.161183 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" containerName="oc" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.161642 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.163581 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.167538 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.175911 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.238320 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.238407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.238497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.340095 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.340168 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.340205 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.341215 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.348768 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.359314 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"collect-profiles-29556885-9xsw6\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.479035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:00 crc kubenswrapper[4907]: I0313 14:45:00.912955 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 14:45:01 crc kubenswrapper[4907]: I0313 14:45:01.451951 4907 generic.go:334] "Generic (PLEG): container finished" podID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerID="34346eaf9fc976e698b78600d581550e67e1b840574df19f7129d1f617293a7d" exitCode=0 Mar 13 14:45:01 crc kubenswrapper[4907]: I0313 14:45:01.451993 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" event={"ID":"d79a5eae-26dc-4f69-aa63-9bd6c5788a03","Type":"ContainerDied","Data":"34346eaf9fc976e698b78600d581550e67e1b840574df19f7129d1f617293a7d"} Mar 13 14:45:01 crc kubenswrapper[4907]: I0313 14:45:01.452017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" event={"ID":"d79a5eae-26dc-4f69-aa63-9bd6c5788a03","Type":"ContainerStarted","Data":"500440f870c9906ee7451693b9eaf24fd71405810b18783bcc0bedd8966b5ea2"} Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.828507 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.878387 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") pod \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.878437 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") pod \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.878507 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") pod \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\" (UID: \"d79a5eae-26dc-4f69-aa63-9bd6c5788a03\") " Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.879565 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume" (OuterVolumeSpecName: "config-volume") pod "d79a5eae-26dc-4f69-aa63-9bd6c5788a03" (UID: "d79a5eae-26dc-4f69-aa63-9bd6c5788a03"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.890237 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d79a5eae-26dc-4f69-aa63-9bd6c5788a03" (UID: "d79a5eae-26dc-4f69-aa63-9bd6c5788a03"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.890591 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb" (OuterVolumeSpecName: "kube-api-access-456sb") pod "d79a5eae-26dc-4f69-aa63-9bd6c5788a03" (UID: "d79a5eae-26dc-4f69-aa63-9bd6c5788a03"). InnerVolumeSpecName "kube-api-access-456sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.980661 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-456sb\" (UniqueName: \"kubernetes.io/projected/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-kube-api-access-456sb\") on node \"crc\" DevicePath \"\"" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.980693 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:45:02 crc kubenswrapper[4907]: I0313 14:45:02.980702 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d79a5eae-26dc-4f69-aa63-9bd6c5788a03-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.471728 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" event={"ID":"d79a5eae-26dc-4f69-aa63-9bd6c5788a03","Type":"ContainerDied","Data":"500440f870c9906ee7451693b9eaf24fd71405810b18783bcc0bedd8966b5ea2"} Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.471787 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="500440f870c9906ee7451693b9eaf24fd71405810b18783bcc0bedd8966b5ea2" Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.471847 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6" Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.914705 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:45:03 crc kubenswrapper[4907]: I0313 14:45:03.922269 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556840-pgmwj"] Mar 13 14:45:05 crc kubenswrapper[4907]: I0313 14:45:05.793098 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19edf37b-a1d3-4c7c-945e-5d0b8bb601bf" path="/var/lib/kubelet/pods/19edf37b-a1d3-4c7c-945e-5d0b8bb601bf/volumes" Mar 13 14:45:07 crc kubenswrapper[4907]: I0313 14:45:07.783410 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:07 crc kubenswrapper[4907]: E0313 14:45:07.784489 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:19 crc kubenswrapper[4907]: I0313 14:45:19.782353 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:19 crc kubenswrapper[4907]: E0313 14:45:19.783056 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:25 crc kubenswrapper[4907]: I0313 14:45:25.136823 4907 scope.go:117] "RemoveContainer" containerID="d2284227e021b37d9e8d2812f8621bebf16efd8eeb5019b51cfa28773e7a5aa9" Mar 13 14:45:31 crc kubenswrapper[4907]: I0313 14:45:31.786706 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:31 crc kubenswrapper[4907]: E0313 14:45:31.787460 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:44 crc kubenswrapper[4907]: I0313 14:45:44.782639 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:44 crc kubenswrapper[4907]: E0313 14:45:44.783580 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:45:58 crc kubenswrapper[4907]: I0313 14:45:58.782384 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:45:58 crc kubenswrapper[4907]: E0313 14:45:58.783092 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.151783 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:46:00 crc kubenswrapper[4907]: E0313 14:46:00.153507 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerName="collect-profiles" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.153625 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerName="collect-profiles" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.153926 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" containerName="collect-profiles" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.154589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.158302 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.159689 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.160484 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.162585 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.223127 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"auto-csr-approver-29556886-tqfp7\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.324175 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"auto-csr-approver-29556886-tqfp7\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.342164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"auto-csr-approver-29556886-tqfp7\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.477541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.916290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:46:00 crc kubenswrapper[4907]: I0313 14:46:00.948589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" event={"ID":"7e68a473-ca00-4a39-bdca-1a122010d02f","Type":"ContainerStarted","Data":"e8d00dfed9d39506586f930b123b132eb8946a22941081dff69f3cd7065de584"} Mar 13 14:46:02 crc kubenswrapper[4907]: I0313 14:46:02.963012 4907 generic.go:334] "Generic (PLEG): container finished" podID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerID="07d959b411798c04dbd7ce18d3d50f7352525fb8aa854268d1f201958d491e3c" exitCode=0 Mar 13 14:46:02 crc kubenswrapper[4907]: I0313 14:46:02.963276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" event={"ID":"7e68a473-ca00-4a39-bdca-1a122010d02f","Type":"ContainerDied","Data":"07d959b411798c04dbd7ce18d3d50f7352525fb8aa854268d1f201958d491e3c"} Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.227117 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.379303 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") pod \"7e68a473-ca00-4a39-bdca-1a122010d02f\" (UID: \"7e68a473-ca00-4a39-bdca-1a122010d02f\") " Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.385627 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj" (OuterVolumeSpecName: "kube-api-access-9skvj") pod "7e68a473-ca00-4a39-bdca-1a122010d02f" (UID: "7e68a473-ca00-4a39-bdca-1a122010d02f"). InnerVolumeSpecName "kube-api-access-9skvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.481999 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9skvj\" (UniqueName: \"kubernetes.io/projected/7e68a473-ca00-4a39-bdca-1a122010d02f-kube-api-access-9skvj\") on node \"crc\" DevicePath \"\"" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.978019 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" event={"ID":"7e68a473-ca00-4a39-bdca-1a122010d02f","Type":"ContainerDied","Data":"e8d00dfed9d39506586f930b123b132eb8946a22941081dff69f3cd7065de584"} Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.978070 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8d00dfed9d39506586f930b123b132eb8946a22941081dff69f3cd7065de584" Mar 13 14:46:04 crc kubenswrapper[4907]: I0313 14:46:04.978175 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556886-tqfp7" Mar 13 14:46:05 crc kubenswrapper[4907]: I0313 14:46:05.291527 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:46:05 crc kubenswrapper[4907]: I0313 14:46:05.298216 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556880-wxdzr"] Mar 13 14:46:05 crc kubenswrapper[4907]: I0313 14:46:05.795046 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20dce4c7-7cbf-4e49-bd0e-93ee53d72551" path="/var/lib/kubelet/pods/20dce4c7-7cbf-4e49-bd0e-93ee53d72551/volumes" Mar 13 14:46:11 crc kubenswrapper[4907]: I0313 14:46:11.788621 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:11 crc kubenswrapper[4907]: E0313 14:46:11.789287 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:25 crc kubenswrapper[4907]: I0313 14:46:25.185072 4907 scope.go:117] "RemoveContainer" containerID="fc46a881237631d2b078045b48ba1243ef9d84385c8036fc8ea4400ece996277" Mar 13 14:46:26 crc kubenswrapper[4907]: I0313 14:46:26.783275 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:26 crc kubenswrapper[4907]: E0313 14:46:26.783787 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:39 crc kubenswrapper[4907]: I0313 14:46:39.782099 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:39 crc kubenswrapper[4907]: E0313 14:46:39.782747 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:46:52 crc kubenswrapper[4907]: I0313 14:46:52.782982 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:46:52 crc kubenswrapper[4907]: E0313 14:46:52.783860 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:07 crc kubenswrapper[4907]: I0313 14:47:07.782550 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:07 crc kubenswrapper[4907]: E0313 14:47:07.783337 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:20 crc kubenswrapper[4907]: I0313 14:47:20.783424 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:20 crc kubenswrapper[4907]: E0313 14:47:20.784704 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:35 crc kubenswrapper[4907]: I0313 14:47:35.782377 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:35 crc kubenswrapper[4907]: E0313 14:47:35.783214 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:47:49 crc kubenswrapper[4907]: I0313 14:47:49.782185 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:47:49 crc kubenswrapper[4907]: E0313 14:47:49.782819 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.168256 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:48:00 crc kubenswrapper[4907]: E0313 14:48:00.169490 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerName="oc" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.169516 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerName="oc" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.169754 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" containerName="oc" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.170491 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.173197 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.173374 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.173809 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.184000 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.334113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"auto-csr-approver-29556888-6dcv4\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.436150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"auto-csr-approver-29556888-6dcv4\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.455014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"auto-csr-approver-29556888-6dcv4\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.498239 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:00 crc kubenswrapper[4907]: I0313 14:48:00.938822 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:48:01 crc kubenswrapper[4907]: I0313 14:48:01.849551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" event={"ID":"4b255770-f854-44fe-93fb-7ae5082081be","Type":"ContainerStarted","Data":"fdf6f7dd09a291140d72d172172a962dbe55255551fbcb81eaab9d7b5b3eb8a5"} Mar 13 14:48:02 crc kubenswrapper[4907]: I0313 14:48:02.859787 4907 generic.go:334] "Generic (PLEG): container finished" podID="4b255770-f854-44fe-93fb-7ae5082081be" containerID="33e7c832124ee1aa4196cb0bd674abcb6224ac6318050e8ed5dbb3afadecce6f" exitCode=0 Mar 13 14:48:02 crc kubenswrapper[4907]: I0313 14:48:02.859843 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" event={"ID":"4b255770-f854-44fe-93fb-7ae5082081be","Type":"ContainerDied","Data":"33e7c832124ee1aa4196cb0bd674abcb6224ac6318050e8ed5dbb3afadecce6f"} Mar 13 14:48:03 crc kubenswrapper[4907]: I0313 14:48:03.783383 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:03 crc kubenswrapper[4907]: E0313 14:48:03.784009 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.135486 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.292582 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") pod \"4b255770-f854-44fe-93fb-7ae5082081be\" (UID: \"4b255770-f854-44fe-93fb-7ae5082081be\") " Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.297608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj" (OuterVolumeSpecName: "kube-api-access-n2nsj") pod "4b255770-f854-44fe-93fb-7ae5082081be" (UID: "4b255770-f854-44fe-93fb-7ae5082081be"). InnerVolumeSpecName "kube-api-access-n2nsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.394575 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2nsj\" (UniqueName: \"kubernetes.io/projected/4b255770-f854-44fe-93fb-7ae5082081be-kube-api-access-n2nsj\") on node \"crc\" DevicePath \"\"" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.881079 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" event={"ID":"4b255770-f854-44fe-93fb-7ae5082081be","Type":"ContainerDied","Data":"fdf6f7dd09a291140d72d172172a962dbe55255551fbcb81eaab9d7b5b3eb8a5"} Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.881156 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdf6f7dd09a291140d72d172172a962dbe55255551fbcb81eaab9d7b5b3eb8a5" Mar 13 14:48:04 crc kubenswrapper[4907]: I0313 14:48:04.881162 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556888-6dcv4" Mar 13 14:48:05 crc kubenswrapper[4907]: I0313 14:48:05.208014 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:48:05 crc kubenswrapper[4907]: I0313 14:48:05.215687 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556882-c7dc8"] Mar 13 14:48:05 crc kubenswrapper[4907]: I0313 14:48:05.798993 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81733636-fc5e-4687-9ceb-0dd2e18fa15a" path="/var/lib/kubelet/pods/81733636-fc5e-4687-9ceb-0dd2e18fa15a/volumes" Mar 13 14:48:17 crc kubenswrapper[4907]: I0313 14:48:17.783576 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:17 crc kubenswrapper[4907]: E0313 14:48:17.784245 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:25 crc kubenswrapper[4907]: I0313 14:48:25.278320 4907 scope.go:117] "RemoveContainer" containerID="888132585d34376894f6b13900aa0293c6191b881a23e9abf82e3ca297b39ec9" Mar 13 14:48:32 crc kubenswrapper[4907]: I0313 14:48:32.782579 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:32 crc kubenswrapper[4907]: E0313 14:48:32.783280 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:44 crc kubenswrapper[4907]: I0313 14:48:44.782256 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:44 crc kubenswrapper[4907]: E0313 14:48:44.782869 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:48:58 crc kubenswrapper[4907]: I0313 14:48:58.782772 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:48:59 crc kubenswrapper[4907]: I0313 14:48:59.324808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352"} Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.145000 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:50:00 crc kubenswrapper[4907]: E0313 14:50:00.146633 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b255770-f854-44fe-93fb-7ae5082081be" containerName="oc" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.146658 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b255770-f854-44fe-93fb-7ae5082081be" containerName="oc" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.146913 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b255770-f854-44fe-93fb-7ae5082081be" containerName="oc" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.147861 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.149944 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.152362 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.152581 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.156187 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.228847 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"auto-csr-approver-29556890-5kp94\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.331031 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"auto-csr-approver-29556890-5kp94\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.352163 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"auto-csr-approver-29556890-5kp94\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.471982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.896038 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:50:00 crc kubenswrapper[4907]: I0313 14:50:00.904753 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:50:01 crc kubenswrapper[4907]: I0313 14:50:01.819020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556890-5kp94" event={"ID":"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40","Type":"ContainerStarted","Data":"9ae7c3fb56c210bef178a4a1f6238ce90b42264a05c6b1a96392a1fe36274965"} Mar 13 14:50:03 crc kubenswrapper[4907]: I0313 14:50:03.840546 4907 generic.go:334] "Generic (PLEG): container finished" podID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerID="f4cc0414786125661ff7aeac8cccd8627fe8bdb42c54577e769f1e1b046242f5" exitCode=0 Mar 13 14:50:03 crc kubenswrapper[4907]: I0313 14:50:03.840623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556890-5kp94" event={"ID":"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40","Type":"ContainerDied","Data":"f4cc0414786125661ff7aeac8cccd8627fe8bdb42c54577e769f1e1b046242f5"} Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.225913 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.309963 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") pod \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\" (UID: \"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40\") " Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.315718 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf" (OuterVolumeSpecName: "kube-api-access-xn2vf") pod "d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" (UID: "d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40"). InnerVolumeSpecName "kube-api-access-xn2vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.411969 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn2vf\" (UniqueName: \"kubernetes.io/projected/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40-kube-api-access-xn2vf\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.856443 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556890-5kp94" event={"ID":"d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40","Type":"ContainerDied","Data":"9ae7c3fb56c210bef178a4a1f6238ce90b42264a05c6b1a96392a1fe36274965"} Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.856751 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ae7c3fb56c210bef178a4a1f6238ce90b42264a05c6b1a96392a1fe36274965" Mar 13 14:50:05 crc kubenswrapper[4907]: I0313 14:50:05.856497 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556890-5kp94" Mar 13 14:50:06 crc kubenswrapper[4907]: I0313 14:50:06.302348 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:50:06 crc kubenswrapper[4907]: I0313 14:50:06.307037 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556884-zzm6t"] Mar 13 14:50:07 crc kubenswrapper[4907]: I0313 14:50:07.793921 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd8a84e4-33ca-4236-8c28-bf80cf64b6bc" path="/var/lib/kubelet/pods/bd8a84e4-33ca-4236-8c28-bf80cf64b6bc/volumes" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.185876 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:15 crc kubenswrapper[4907]: E0313 14:50:15.186671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerName="oc" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.186682 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerName="oc" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.186813 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" containerName="oc" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.187806 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.192506 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.266804 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.267138 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.267242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.368247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.368813 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.369512 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.370159 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.370192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.389529 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"redhat-marketplace-tbj2h\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.502927 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.916488 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:15 crc kubenswrapper[4907]: I0313 14:50:15.938774 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerStarted","Data":"a49953d7498cf6c74e49e1a76f262c8cf7a2a5083aa06bb1bb21fd359058fb8f"} Mar 13 14:50:16 crc kubenswrapper[4907]: I0313 14:50:16.949643 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" exitCode=0 Mar 13 14:50:16 crc kubenswrapper[4907]: I0313 14:50:16.949708 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1"} Mar 13 14:50:17 crc kubenswrapper[4907]: I0313 14:50:17.957237 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" exitCode=0 Mar 13 14:50:17 crc kubenswrapper[4907]: I0313 14:50:17.957277 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0"} Mar 13 14:50:18 crc kubenswrapper[4907]: I0313 14:50:18.965619 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerStarted","Data":"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12"} Mar 13 14:50:18 crc kubenswrapper[4907]: I0313 14:50:18.986661 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tbj2h" podStartSLOduration=2.576377226 podStartE2EDuration="3.986644117s" podCreationTimestamp="2026-03-13 14:50:15 +0000 UTC" firstStartedPulling="2026-03-13 14:50:16.953796667 +0000 UTC m=+2715.853584356" lastFinishedPulling="2026-03-13 14:50:18.364063518 +0000 UTC m=+2717.263851247" observedRunningTime="2026-03-13 14:50:18.982912665 +0000 UTC m=+2717.882700364" watchObservedRunningTime="2026-03-13 14:50:18.986644117 +0000 UTC m=+2717.886431806" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.388808 4907 scope.go:117] "RemoveContainer" containerID="1d99f9504e5b5c3d1b261df414bc1841776b381d4074946ac4e7f45b3d386557" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.503945 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.504557 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:25 crc kubenswrapper[4907]: I0313 14:50:25.545011 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:26 crc kubenswrapper[4907]: I0313 14:50:26.054306 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:26 crc kubenswrapper[4907]: I0313 14:50:26.096250 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.036466 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tbj2h" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" containerID="cri-o://88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" gracePeriod=2 Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.406285 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.535449 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") pod \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.535595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") pod \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.535677 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") pod \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\" (UID: \"b7f6fd9e-d07b-4339-97aa-50903ea7b614\") " Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.536969 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities" (OuterVolumeSpecName: "utilities") pod "b7f6fd9e-d07b-4339-97aa-50903ea7b614" (UID: "b7f6fd9e-d07b-4339-97aa-50903ea7b614"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.542729 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz" (OuterVolumeSpecName: "kube-api-access-kgqmz") pod "b7f6fd9e-d07b-4339-97aa-50903ea7b614" (UID: "b7f6fd9e-d07b-4339-97aa-50903ea7b614"). InnerVolumeSpecName "kube-api-access-kgqmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.571427 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7f6fd9e-d07b-4339-97aa-50903ea7b614" (UID: "b7f6fd9e-d07b-4339-97aa-50903ea7b614"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.637701 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgqmz\" (UniqueName: \"kubernetes.io/projected/b7f6fd9e-d07b-4339-97aa-50903ea7b614-kube-api-access-kgqmz\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.637732 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:28 crc kubenswrapper[4907]: I0313 14:50:28.637742 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7f6fd9e-d07b-4339-97aa-50903ea7b614-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.047061 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tbj2h" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.047102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12"} Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.047231 4907 scope.go:117] "RemoveContainer" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.046845 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" exitCode=0 Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.056103 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tbj2h" event={"ID":"b7f6fd9e-d07b-4339-97aa-50903ea7b614","Type":"ContainerDied","Data":"a49953d7498cf6c74e49e1a76f262c8cf7a2a5083aa06bb1bb21fd359058fb8f"} Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.084007 4907 scope.go:117] "RemoveContainer" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.090534 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.102775 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tbj2h"] Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.112047 4907 scope.go:117] "RemoveContainer" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.136613 4907 scope.go:117] "RemoveContainer" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" Mar 13 14:50:29 crc kubenswrapper[4907]: E0313 14:50:29.137196 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12\": container with ID starting with 88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12 not found: ID does not exist" containerID="88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.137240 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12"} err="failed to get container status \"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12\": rpc error: code = NotFound desc = could not find container \"88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12\": container with ID starting with 88decf377cd617e46acde155df67a278293b44791d1045d2c952e59fa72f3b12 not found: ID does not exist" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.137268 4907 scope.go:117] "RemoveContainer" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" Mar 13 14:50:29 crc kubenswrapper[4907]: E0313 14:50:29.138231 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0\": container with ID starting with 7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0 not found: ID does not exist" containerID="7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.138256 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0"} err="failed to get container status \"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0\": rpc error: code = NotFound desc = could not find container \"7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0\": container with ID starting with 7e5377b54b517be40da3dfb2c0604762de65b6f2cdabc2cfb353510ffcd39ab0 not found: ID does not exist" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.138270 4907 scope.go:117] "RemoveContainer" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" Mar 13 14:50:29 crc kubenswrapper[4907]: E0313 14:50:29.138706 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1\": container with ID starting with be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1 not found: ID does not exist" containerID="be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.138744 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1"} err="failed to get container status \"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1\": rpc error: code = NotFound desc = could not find container \"be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1\": container with ID starting with be8896282b4c21bb9015035503186642512b3e9f125e62a6c4d4cf624108a9f1 not found: ID does not exist" Mar 13 14:50:29 crc kubenswrapper[4907]: I0313 14:50:29.793653 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" path="/var/lib/kubelet/pods/b7f6fd9e-d07b-4339-97aa-50903ea7b614/volumes" Mar 13 14:51:18 crc kubenswrapper[4907]: I0313 14:51:18.041500 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:51:18 crc kubenswrapper[4907]: I0313 14:51:18.042096 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:51:48 crc kubenswrapper[4907]: I0313 14:51:48.041255 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:51:48 crc kubenswrapper[4907]: I0313 14:51:48.041875 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.151860 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:00 crc kubenswrapper[4907]: E0313 14:52:00.152723 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152736 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" Mar 13 14:52:00 crc kubenswrapper[4907]: E0313 14:52:00.152766 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-utilities" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152773 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-utilities" Mar 13 14:52:00 crc kubenswrapper[4907]: E0313 14:52:00.152786 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-content" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152792 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="extract-content" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.152940 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f6fd9e-d07b-4339-97aa-50903ea7b614" containerName="registry-server" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.153870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.167549 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.244372 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.244458 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.244801 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.251463 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.258943 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.260567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.261793 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.263166 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.264241 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.346828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.346961 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"auto-csr-approver-29556892-gq5qv\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.346997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.347023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.347495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.347547 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.367841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"certified-operators-hglxf\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.448856 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"auto-csr-approver-29556892-gq5qv\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.466994 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"auto-csr-approver-29556892-gq5qv\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.474925 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.575306 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:00 crc kubenswrapper[4907]: I0313 14:52:00.939988 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:01 crc kubenswrapper[4907]: I0313 14:52:01.069426 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:52:01 crc kubenswrapper[4907]: I0313 14:52:01.072290 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerStarted","Data":"76918a7b1d9b781c78870a05cada882f98b668b545bb9ef8a4888d25a3d213a8"} Mar 13 14:52:01 crc kubenswrapper[4907]: W0313 14:52:01.084217 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbb8b9ec_85bd_4617_952d_57ce0bdb4553.slice/crio-6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa WatchSource:0}: Error finding container 6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa: Status 404 returned error can't find the container with id 6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa Mar 13 14:52:02 crc kubenswrapper[4907]: I0313 14:52:02.080694 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" event={"ID":"cbb8b9ec-85bd-4617-952d-57ce0bdb4553","Type":"ContainerStarted","Data":"6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa"} Mar 13 14:52:02 crc kubenswrapper[4907]: I0313 14:52:02.082504 4907 generic.go:334] "Generic (PLEG): container finished" podID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" exitCode=0 Mar 13 14:52:02 crc kubenswrapper[4907]: I0313 14:52:02.082859 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c"} Mar 13 14:52:03 crc kubenswrapper[4907]: I0313 14:52:03.090237 4907 generic.go:334] "Generic (PLEG): container finished" podID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerID="2c73e4b83765c168348350266e6ee7d5565c27f33d1acbef663a50d336ac5e75" exitCode=0 Mar 13 14:52:03 crc kubenswrapper[4907]: I0313 14:52:03.090469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" event={"ID":"cbb8b9ec-85bd-4617-952d-57ce0bdb4553","Type":"ContainerDied","Data":"2c73e4b83765c168348350266e6ee7d5565c27f33d1acbef663a50d336ac5e75"} Mar 13 14:52:03 crc kubenswrapper[4907]: I0313 14:52:03.093271 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerStarted","Data":"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8"} Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.101856 4907 generic.go:334] "Generic (PLEG): container finished" podID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" exitCode=0 Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.101912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8"} Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.405837 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.504952 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") pod \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\" (UID: \"cbb8b9ec-85bd-4617-952d-57ce0bdb4553\") " Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.511675 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg" (OuterVolumeSpecName: "kube-api-access-k8zbg") pod "cbb8b9ec-85bd-4617-952d-57ce0bdb4553" (UID: "cbb8b9ec-85bd-4617-952d-57ce0bdb4553"). InnerVolumeSpecName "kube-api-access-k8zbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:52:04 crc kubenswrapper[4907]: I0313 14:52:04.606456 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8zbg\" (UniqueName: \"kubernetes.io/projected/cbb8b9ec-85bd-4617-952d-57ce0bdb4553-kube-api-access-k8zbg\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.110320 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.110323 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556892-gq5qv" event={"ID":"cbb8b9ec-85bd-4617-952d-57ce0bdb4553","Type":"ContainerDied","Data":"6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa"} Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.110693 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6540b5f3ac658f051ef622c91f789b6c6c1726d92e07753b794ac5b4496497aa" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.112219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerStarted","Data":"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e"} Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.136316 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hglxf" podStartSLOduration=2.495860144 podStartE2EDuration="5.136298253s" podCreationTimestamp="2026-03-13 14:52:00 +0000 UTC" firstStartedPulling="2026-03-13 14:52:02.084588612 +0000 UTC m=+2820.984376301" lastFinishedPulling="2026-03-13 14:52:04.725026721 +0000 UTC m=+2823.624814410" observedRunningTime="2026-03-13 14:52:05.132686735 +0000 UTC m=+2824.032474424" watchObservedRunningTime="2026-03-13 14:52:05.136298253 +0000 UTC m=+2824.036085942" Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.462356 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.468462 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556886-tqfp7"] Mar 13 14:52:05 crc kubenswrapper[4907]: I0313 14:52:05.790146 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e68a473-ca00-4a39-bdca-1a122010d02f" path="/var/lib/kubelet/pods/7e68a473-ca00-4a39-bdca-1a122010d02f/volumes" Mar 13 14:52:10 crc kubenswrapper[4907]: I0313 14:52:10.475553 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:10 crc kubenswrapper[4907]: I0313 14:52:10.476272 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:10 crc kubenswrapper[4907]: I0313 14:52:10.521268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:11 crc kubenswrapper[4907]: I0313 14:52:11.200294 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:11 crc kubenswrapper[4907]: I0313 14:52:11.250863 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.173327 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hglxf" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" containerID="cri-o://08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" gracePeriod=2 Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.691548 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.736752 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") pod \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.736862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") pod \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.736990 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") pod \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\" (UID: \"73e2ff11-5a7d-4fd9-9908-b57c774d4793\") " Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.738115 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities" (OuterVolumeSpecName: "utilities") pod "73e2ff11-5a7d-4fd9-9908-b57c774d4793" (UID: "73e2ff11-5a7d-4fd9-9908-b57c774d4793"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.744196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf" (OuterVolumeSpecName: "kube-api-access-jhsvf") pod "73e2ff11-5a7d-4fd9-9908-b57c774d4793" (UID: "73e2ff11-5a7d-4fd9-9908-b57c774d4793"). InnerVolumeSpecName "kube-api-access-jhsvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.809342 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73e2ff11-5a7d-4fd9-9908-b57c774d4793" (UID: "73e2ff11-5a7d-4fd9-9908-b57c774d4793"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.838970 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.839004 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhsvf\" (UniqueName: \"kubernetes.io/projected/73e2ff11-5a7d-4fd9-9908-b57c774d4793-kube-api-access-jhsvf\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:13 crc kubenswrapper[4907]: I0313 14:52:13.839017 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73e2ff11-5a7d-4fd9-9908-b57c774d4793-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185695 4907 generic.go:334] "Generic (PLEG): container finished" podID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" exitCode=0 Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185783 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e"} Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185839 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hglxf" event={"ID":"73e2ff11-5a7d-4fd9-9908-b57c774d4793","Type":"ContainerDied","Data":"76918a7b1d9b781c78870a05cada882f98b668b545bb9ef8a4888d25a3d213a8"} Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.185923 4907 scope.go:117] "RemoveContainer" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.186171 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hglxf" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.217835 4907 scope.go:117] "RemoveContainer" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.231201 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.244090 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hglxf"] Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.248147 4907 scope.go:117] "RemoveContainer" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.269504 4907 scope.go:117] "RemoveContainer" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" Mar 13 14:52:14 crc kubenswrapper[4907]: E0313 14:52:14.270009 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e\": container with ID starting with 08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e not found: ID does not exist" containerID="08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270059 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e"} err="failed to get container status \"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e\": rpc error: code = NotFound desc = could not find container \"08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e\": container with ID starting with 08451a948422d2c35b846b2a876db014a9cf925cdae25fd1738e4b8845d5de4e not found: ID does not exist" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270092 4907 scope.go:117] "RemoveContainer" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" Mar 13 14:52:14 crc kubenswrapper[4907]: E0313 14:52:14.270531 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8\": container with ID starting with 69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8 not found: ID does not exist" containerID="69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270843 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8"} err="failed to get container status \"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8\": rpc error: code = NotFound desc = could not find container \"69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8\": container with ID starting with 69543302314ed4849971909b44e1e718b46042e5aea85023f93ed2a8109086e8 not found: ID does not exist" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.270910 4907 scope.go:117] "RemoveContainer" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" Mar 13 14:52:14 crc kubenswrapper[4907]: E0313 14:52:14.271251 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c\": container with ID starting with 1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c not found: ID does not exist" containerID="1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c" Mar 13 14:52:14 crc kubenswrapper[4907]: I0313 14:52:14.271289 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c"} err="failed to get container status \"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c\": rpc error: code = NotFound desc = could not find container \"1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c\": container with ID starting with 1039a932888e0112ca2f48289a4503fb0f292131eefcd895d4539232ce58909c not found: ID does not exist" Mar 13 14:52:15 crc kubenswrapper[4907]: I0313 14:52:15.811089 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" path="/var/lib/kubelet/pods/73e2ff11-5a7d-4fd9-9908-b57c774d4793/volumes" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.041728 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042085 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042167 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042865 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.042958 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352" gracePeriod=600 Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.228785 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352" exitCode=0 Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.228866 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352"} Mar 13 14:52:18 crc kubenswrapper[4907]: I0313 14:52:18.228992 4907 scope.go:117] "RemoveContainer" containerID="43810caaacd8c4e6d986aec4d7996e802ba4907ce25cbd96e3f26a101537186b" Mar 13 14:52:19 crc kubenswrapper[4907]: I0313 14:52:19.240103 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3"} Mar 13 14:52:25 crc kubenswrapper[4907]: I0313 14:52:25.472983 4907 scope.go:117] "RemoveContainer" containerID="07d959b411798c04dbd7ce18d3d50f7352525fb8aa854268d1f201958d491e3c" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.334772 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335667 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerName="oc" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335683 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerName="oc" Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335697 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-content" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335718 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-content" Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335739 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-utilities" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335748 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="extract-utilities" Mar 13 14:52:31 crc kubenswrapper[4907]: E0313 14:52:31.335759 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335767 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335944 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" containerName="oc" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.335978 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="73e2ff11-5a7d-4fd9-9908-b57c774d4793" containerName="registry-server" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.337148 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.349540 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.478245 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.478592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.478804 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.579693 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.579750 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.579779 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.580266 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.580352 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.609338 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"redhat-operators-s8srt\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:31 crc kubenswrapper[4907]: I0313 14:52:31.657689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.088895 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:32 crc kubenswrapper[4907]: W0313 14:52:32.100538 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda53a97c4_ae1d_4e67_b816_c1e4852b08e5.slice/crio-0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715 WatchSource:0}: Error finding container 0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715: Status 404 returned error can't find the container with id 0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715 Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.331160 4907 generic.go:334] "Generic (PLEG): container finished" podID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerID="0b057665a0d00f3a48117344966f922b2928ae4d7a2f92b272a3107b565359be" exitCode=0 Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.331245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"0b057665a0d00f3a48117344966f922b2928ae4d7a2f92b272a3107b565359be"} Mar 13 14:52:32 crc kubenswrapper[4907]: I0313 14:52:32.331855 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerStarted","Data":"0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715"} Mar 13 14:52:33 crc kubenswrapper[4907]: I0313 14:52:33.340476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerStarted","Data":"5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920"} Mar 13 14:52:34 crc kubenswrapper[4907]: I0313 14:52:34.348772 4907 generic.go:334] "Generic (PLEG): container finished" podID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerID="5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920" exitCode=0 Mar 13 14:52:34 crc kubenswrapper[4907]: I0313 14:52:34.348829 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920"} Mar 13 14:52:36 crc kubenswrapper[4907]: I0313 14:52:36.364793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerStarted","Data":"a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d"} Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.658611 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.658979 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.700549 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:41 crc kubenswrapper[4907]: I0313 14:52:41.721655 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s8srt" podStartSLOduration=7.649174469 podStartE2EDuration="10.721639457s" podCreationTimestamp="2026-03-13 14:52:31 +0000 UTC" firstStartedPulling="2026-03-13 14:52:32.332464934 +0000 UTC m=+2851.232252623" lastFinishedPulling="2026-03-13 14:52:35.404929922 +0000 UTC m=+2854.304717611" observedRunningTime="2026-03-13 14:52:36.384110369 +0000 UTC m=+2855.283898078" watchObservedRunningTime="2026-03-13 14:52:41.721639457 +0000 UTC m=+2860.621427146" Mar 13 14:52:42 crc kubenswrapper[4907]: I0313 14:52:42.441830 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:42 crc kubenswrapper[4907]: I0313 14:52:42.494762 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:44 crc kubenswrapper[4907]: I0313 14:52:44.414463 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s8srt" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" containerID="cri-o://a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d" gracePeriod=2 Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.422717 4907 generic.go:334] "Generic (PLEG): container finished" podID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerID="a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d" exitCode=0 Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.422787 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d"} Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.927559 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.976270 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") pod \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " Mar 13 14:52:45 crc kubenswrapper[4907]: I0313 14:52:45.982087 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz" (OuterVolumeSpecName: "kube-api-access-892qz") pod "a53a97c4-ae1d-4e67-b816-c1e4852b08e5" (UID: "a53a97c4-ae1d-4e67-b816-c1e4852b08e5"). InnerVolumeSpecName "kube-api-access-892qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.077829 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") pod \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.077938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") pod \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\" (UID: \"a53a97c4-ae1d-4e67-b816-c1e4852b08e5\") " Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.078297 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-892qz\" (UniqueName: \"kubernetes.io/projected/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-kube-api-access-892qz\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.079105 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities" (OuterVolumeSpecName: "utilities") pod "a53a97c4-ae1d-4e67-b816-c1e4852b08e5" (UID: "a53a97c4-ae1d-4e67-b816-c1e4852b08e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.179526 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.210780 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a53a97c4-ae1d-4e67-b816-c1e4852b08e5" (UID: "a53a97c4-ae1d-4e67-b816-c1e4852b08e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.281546 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a53a97c4-ae1d-4e67-b816-c1e4852b08e5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.435133 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s8srt" event={"ID":"a53a97c4-ae1d-4e67-b816-c1e4852b08e5","Type":"ContainerDied","Data":"0fd20c97ede4d0b2cf316bf01e48438f62ec92b557262bb0310559909098c715"} Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.435207 4907 scope.go:117] "RemoveContainer" containerID="a0f45585bd9bea343e63d05743c65bf9a3e6218aa8b37cbe283fae944e8b7e8d" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.435280 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s8srt" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.457101 4907 scope.go:117] "RemoveContainer" containerID="5f1bfe1233a28e84068fa3690e2155ac3883d0d7f36e3aac56ab833ab2614920" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.481158 4907 scope.go:117] "RemoveContainer" containerID="0b057665a0d00f3a48117344966f922b2928ae4d7a2f92b272a3107b565359be" Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.497168 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:46 crc kubenswrapper[4907]: I0313 14:52:46.504347 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s8srt"] Mar 13 14:52:47 crc kubenswrapper[4907]: I0313 14:52:47.790600 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" path="/var/lib/kubelet/pods/a53a97c4-ae1d-4e67-b816-c1e4852b08e5/volumes" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.557327 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:40 crc kubenswrapper[4907]: E0313 14:53:40.558105 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558117 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" Mar 13 14:53:40 crc kubenswrapper[4907]: E0313 14:53:40.558137 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-content" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558142 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-content" Mar 13 14:53:40 crc kubenswrapper[4907]: E0313 14:53:40.558165 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-utilities" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558172 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="extract-utilities" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.558298 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a53a97c4-ae1d-4e67-b816-c1e4852b08e5" containerName="registry-server" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.559230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.575687 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.579965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.580061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.580124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681246 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681351 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681812 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.681860 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.711667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"community-operators-zpk2j\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:40 crc kubenswrapper[4907]: I0313 14:53:40.877731 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.346861 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.832227 4907 generic.go:334] "Generic (PLEG): container finished" podID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" exitCode=0 Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.832295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850"} Mar 13 14:53:41 crc kubenswrapper[4907]: I0313 14:53:41.832471 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerStarted","Data":"507df3aaee8a6dbfeb22b5e97bee364fc1ad89fcaba6062646ae95dc721d6f17"} Mar 13 14:53:43 crc kubenswrapper[4907]: I0313 14:53:43.847154 4907 generic.go:334] "Generic (PLEG): container finished" podID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" exitCode=0 Mar 13 14:53:43 crc kubenswrapper[4907]: I0313 14:53:43.847215 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67"} Mar 13 14:53:44 crc kubenswrapper[4907]: I0313 14:53:44.857066 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerStarted","Data":"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8"} Mar 13 14:53:44 crc kubenswrapper[4907]: I0313 14:53:44.875546 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zpk2j" podStartSLOduration=2.430917728 podStartE2EDuration="4.875529439s" podCreationTimestamp="2026-03-13 14:53:40 +0000 UTC" firstStartedPulling="2026-03-13 14:53:41.834489641 +0000 UTC m=+2920.734277330" lastFinishedPulling="2026-03-13 14:53:44.279101312 +0000 UTC m=+2923.178889041" observedRunningTime="2026-03-13 14:53:44.872452284 +0000 UTC m=+2923.772239973" watchObservedRunningTime="2026-03-13 14:53:44.875529439 +0000 UTC m=+2923.775317128" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.878767 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.880085 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.926631 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:50 crc kubenswrapper[4907]: I0313 14:53:50.969747 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:51 crc kubenswrapper[4907]: I0313 14:53:51.508944 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:52 crc kubenswrapper[4907]: I0313 14:53:52.913814 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zpk2j" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" containerID="cri-o://0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" gracePeriod=2 Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.270557 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.464669 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") pod \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.464789 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") pod \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.464818 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") pod \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\" (UID: \"40a01f10-ea81-4843-bee8-cc7e5851b1e0\") " Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.466221 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities" (OuterVolumeSpecName: "utilities") pod "40a01f10-ea81-4843-bee8-cc7e5851b1e0" (UID: "40a01f10-ea81-4843-bee8-cc7e5851b1e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.471257 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2" (OuterVolumeSpecName: "kube-api-access-blcs2") pod "40a01f10-ea81-4843-bee8-cc7e5851b1e0" (UID: "40a01f10-ea81-4843-bee8-cc7e5851b1e0"). InnerVolumeSpecName "kube-api-access-blcs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.523409 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40a01f10-ea81-4843-bee8-cc7e5851b1e0" (UID: "40a01f10-ea81-4843-bee8-cc7e5851b1e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.572358 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.572405 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a01f10-ea81-4843-bee8-cc7e5851b1e0-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.572420 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blcs2\" (UniqueName: \"kubernetes.io/projected/40a01f10-ea81-4843-bee8-cc7e5851b1e0-kube-api-access-blcs2\") on node \"crc\" DevicePath \"\"" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922300 4907 generic.go:334] "Generic (PLEG): container finished" podID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" exitCode=0 Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922359 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8"} Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zpk2j" event={"ID":"40a01f10-ea81-4843-bee8-cc7e5851b1e0","Type":"ContainerDied","Data":"507df3aaee8a6dbfeb22b5e97bee364fc1ad89fcaba6062646ae95dc721d6f17"} Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922364 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zpk2j" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.922442 4907 scope.go:117] "RemoveContainer" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.945316 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.946565 4907 scope.go:117] "RemoveContainer" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.950636 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zpk2j"] Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.965708 4907 scope.go:117] "RemoveContainer" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.981576 4907 scope.go:117] "RemoveContainer" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" Mar 13 14:53:53 crc kubenswrapper[4907]: E0313 14:53:53.981928 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8\": container with ID starting with 0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8 not found: ID does not exist" containerID="0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.981965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8"} err="failed to get container status \"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8\": rpc error: code = NotFound desc = could not find container \"0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8\": container with ID starting with 0d00e6e05c6219b1e79b7b2d7b334c8b87d4b7aeb54d70b84bcf93cbd75db2d8 not found: ID does not exist" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982006 4907 scope.go:117] "RemoveContainer" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" Mar 13 14:53:53 crc kubenswrapper[4907]: E0313 14:53:53.982226 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67\": container with ID starting with 401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67 not found: ID does not exist" containerID="401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982252 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67"} err="failed to get container status \"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67\": rpc error: code = NotFound desc = could not find container \"401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67\": container with ID starting with 401a1d6750944d46aa81b47dac6e65af576075e988be53e4b38e87b67ecebd67 not found: ID does not exist" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982266 4907 scope.go:117] "RemoveContainer" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" Mar 13 14:53:53 crc kubenswrapper[4907]: E0313 14:53:53.982570 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850\": container with ID starting with 90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850 not found: ID does not exist" containerID="90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850" Mar 13 14:53:53 crc kubenswrapper[4907]: I0313 14:53:53.982590 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850"} err="failed to get container status \"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850\": rpc error: code = NotFound desc = could not find container \"90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850\": container with ID starting with 90a60a553cdd587785b5c5a14302529824eeffe1f263fc6e5179be6c780af850 not found: ID does not exist" Mar 13 14:53:55 crc kubenswrapper[4907]: I0313 14:53:55.792635 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" path="/var/lib/kubelet/pods/40a01f10-ea81-4843-bee8-cc7e5851b1e0/volumes" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.140385 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 14:54:00 crc kubenswrapper[4907]: E0313 14:54:00.141250 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-content" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141266 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-content" Mar 13 14:54:00 crc kubenswrapper[4907]: E0313 14:54:00.141283 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141293 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" Mar 13 14:54:00 crc kubenswrapper[4907]: E0313 14:54:00.141310 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-utilities" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141319 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="extract-utilities" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.141509 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="40a01f10-ea81-4843-bee8-cc7e5851b1e0" containerName="registry-server" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.142120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.144540 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.145744 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.150996 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.153208 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.161570 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"auto-csr-approver-29556894-8mk4g\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.262680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"auto-csr-approver-29556894-8mk4g\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.281011 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"auto-csr-approver-29556894-8mk4g\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.458983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.884993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 14:54:00 crc kubenswrapper[4907]: I0313 14:54:00.979589 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" event={"ID":"14af3ced-cea4-46f4-82b2-bdcc55744a4f","Type":"ContainerStarted","Data":"6c0e1e2490d9f0a95c8081e3db1bb9df146ef5a4b0b734e629bd40b26691c155"} Mar 13 14:54:03 crc kubenswrapper[4907]: I0313 14:54:03.000040 4907 generic.go:334] "Generic (PLEG): container finished" podID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerID="1f3fec4f332e7900eb7076ad444f074ae92511e972577dd77d05eebc0931b626" exitCode=0 Mar 13 14:54:03 crc kubenswrapper[4907]: I0313 14:54:03.000123 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" event={"ID":"14af3ced-cea4-46f4-82b2-bdcc55744a4f","Type":"ContainerDied","Data":"1f3fec4f332e7900eb7076ad444f074ae92511e972577dd77d05eebc0931b626"} Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.434050 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.629861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") pod \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\" (UID: \"14af3ced-cea4-46f4-82b2-bdcc55744a4f\") " Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.634581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8" (OuterVolumeSpecName: "kube-api-access-dkvf8") pod "14af3ced-cea4-46f4-82b2-bdcc55744a4f" (UID: "14af3ced-cea4-46f4-82b2-bdcc55744a4f"). InnerVolumeSpecName "kube-api-access-dkvf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:54:04 crc kubenswrapper[4907]: I0313 14:54:04.732001 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkvf8\" (UniqueName: \"kubernetes.io/projected/14af3ced-cea4-46f4-82b2-bdcc55744a4f-kube-api-access-dkvf8\") on node \"crc\" DevicePath \"\"" Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.016189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" event={"ID":"14af3ced-cea4-46f4-82b2-bdcc55744a4f","Type":"ContainerDied","Data":"6c0e1e2490d9f0a95c8081e3db1bb9df146ef5a4b0b734e629bd40b26691c155"} Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.016239 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c0e1e2490d9f0a95c8081e3db1bb9df146ef5a4b0b734e629bd40b26691c155" Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.016244 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556894-8mk4g" Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.500023 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.508711 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556888-6dcv4"] Mar 13 14:54:05 crc kubenswrapper[4907]: I0313 14:54:05.795166 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b255770-f854-44fe-93fb-7ae5082081be" path="/var/lib/kubelet/pods/4b255770-f854-44fe-93fb-7ae5082081be/volumes" Mar 13 14:54:18 crc kubenswrapper[4907]: I0313 14:54:18.041849 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:54:18 crc kubenswrapper[4907]: I0313 14:54:18.042337 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:54:25 crc kubenswrapper[4907]: I0313 14:54:25.574570 4907 scope.go:117] "RemoveContainer" containerID="33e7c832124ee1aa4196cb0bd674abcb6224ac6318050e8ed5dbb3afadecce6f" Mar 13 14:54:48 crc kubenswrapper[4907]: I0313 14:54:48.042033 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:54:48 crc kubenswrapper[4907]: I0313 14:54:48.042606 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.041927 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.042512 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.042566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.043268 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.043344 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" gracePeriod=600 Mar 13 14:55:18 crc kubenswrapper[4907]: E0313 14:55:18.167225 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.562674 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" exitCode=0 Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.562738 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3"} Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.562788 4907 scope.go:117] "RemoveContainer" containerID="5b16948da306cc82d0f15e426031c3bf4ecc434619902cfc2cfc82bb26ad6352" Mar 13 14:55:18 crc kubenswrapper[4907]: I0313 14:55:18.563647 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:18 crc kubenswrapper[4907]: E0313 14:55:18.564081 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:31 crc kubenswrapper[4907]: I0313 14:55:31.790212 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:31 crc kubenswrapper[4907]: E0313 14:55:31.791402 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:42 crc kubenswrapper[4907]: I0313 14:55:42.782798 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:42 crc kubenswrapper[4907]: E0313 14:55:42.783592 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:55:53 crc kubenswrapper[4907]: I0313 14:55:53.782759 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:55:53 crc kubenswrapper[4907]: E0313 14:55:53.784223 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.144802 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 14:56:00 crc kubenswrapper[4907]: E0313 14:56:00.145804 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerName="oc" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.145826 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerName="oc" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.146117 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" containerName="oc" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.146789 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.151150 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.151483 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.151650 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.159437 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.316399 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"auto-csr-approver-29556896-7dlx2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.417726 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"auto-csr-approver-29556896-7dlx2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.448853 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"auto-csr-approver-29556896-7dlx2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.471979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.911827 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.923212 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 14:56:00 crc kubenswrapper[4907]: I0313 14:56:00.936952 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" event={"ID":"e193f9f5-a033-4f42-9932-62f7b4edd0b2","Type":"ContainerStarted","Data":"fe1ac4052638fbf8e2ed7b02b6cc93e68c5a8f741a2e3a0784fd85d47094c7ff"} Mar 13 14:56:02 crc kubenswrapper[4907]: I0313 14:56:02.954006 4907 generic.go:334] "Generic (PLEG): container finished" podID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerID="19c525d3167fb54f80c181f5117cf495aa5c0602e81951a3395f3a05b6fbc6c9" exitCode=0 Mar 13 14:56:02 crc kubenswrapper[4907]: I0313 14:56:02.954140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" event={"ID":"e193f9f5-a033-4f42-9932-62f7b4edd0b2","Type":"ContainerDied","Data":"19c525d3167fb54f80c181f5117cf495aa5c0602e81951a3395f3a05b6fbc6c9"} Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.264171 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.283110 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") pod \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\" (UID: \"e193f9f5-a033-4f42-9932-62f7b4edd0b2\") " Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.288556 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv" (OuterVolumeSpecName: "kube-api-access-j9wkv") pod "e193f9f5-a033-4f42-9932-62f7b4edd0b2" (UID: "e193f9f5-a033-4f42-9932-62f7b4edd0b2"). InnerVolumeSpecName "kube-api-access-j9wkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.384497 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9wkv\" (UniqueName: \"kubernetes.io/projected/e193f9f5-a033-4f42-9932-62f7b4edd0b2-kube-api-access-j9wkv\") on node \"crc\" DevicePath \"\"" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.971768 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" event={"ID":"e193f9f5-a033-4f42-9932-62f7b4edd0b2","Type":"ContainerDied","Data":"fe1ac4052638fbf8e2ed7b02b6cc93e68c5a8f741a2e3a0784fd85d47094c7ff"} Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.971808 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe1ac4052638fbf8e2ed7b02b6cc93e68c5a8f741a2e3a0784fd85d47094c7ff" Mar 13 14:56:04 crc kubenswrapper[4907]: I0313 14:56:04.971856 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556896-7dlx2" Mar 13 14:56:05 crc kubenswrapper[4907]: I0313 14:56:05.329744 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:56:05 crc kubenswrapper[4907]: I0313 14:56:05.335771 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556890-5kp94"] Mar 13 14:56:05 crc kubenswrapper[4907]: I0313 14:56:05.790828 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40" path="/var/lib/kubelet/pods/d63f291e-b12e-4b0a-b4c3-e7b9c2cc3c40/volumes" Mar 13 14:56:07 crc kubenswrapper[4907]: I0313 14:56:07.782526 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:07 crc kubenswrapper[4907]: E0313 14:56:07.783101 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:20 crc kubenswrapper[4907]: I0313 14:56:20.782968 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:20 crc kubenswrapper[4907]: E0313 14:56:20.783796 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:25 crc kubenswrapper[4907]: I0313 14:56:25.714329 4907 scope.go:117] "RemoveContainer" containerID="f4cc0414786125661ff7aeac8cccd8627fe8bdb42c54577e769f1e1b046242f5" Mar 13 14:56:34 crc kubenswrapper[4907]: I0313 14:56:34.782733 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:34 crc kubenswrapper[4907]: E0313 14:56:34.783515 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:56:47 crc kubenswrapper[4907]: I0313 14:56:47.950910 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:56:47 crc kubenswrapper[4907]: E0313 14:56:47.951630 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:01 crc kubenswrapper[4907]: I0313 14:57:01.787123 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:01 crc kubenswrapper[4907]: E0313 14:57:01.788053 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:13 crc kubenswrapper[4907]: I0313 14:57:13.783488 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:13 crc kubenswrapper[4907]: E0313 14:57:13.784646 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:24 crc kubenswrapper[4907]: I0313 14:57:24.783355 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:24 crc kubenswrapper[4907]: E0313 14:57:24.784512 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:36 crc kubenswrapper[4907]: I0313 14:57:36.782498 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:36 crc kubenswrapper[4907]: E0313 14:57:36.784535 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:57:48 crc kubenswrapper[4907]: I0313 14:57:48.783019 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:57:48 crc kubenswrapper[4907]: E0313 14:57:48.783684 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.140246 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 14:58:00 crc kubenswrapper[4907]: E0313 14:58:00.141141 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerName="oc" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.141156 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerName="oc" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.141340 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" containerName="oc" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.141871 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.144152 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.144941 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.148774 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.150908 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.195098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"auto-csr-approver-29556898-55zxj\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.296587 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"auto-csr-approver-29556898-55zxj\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.324422 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"auto-csr-approver-29556898-55zxj\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.461985 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:00 crc kubenswrapper[4907]: I0313 14:58:00.874638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 14:58:01 crc kubenswrapper[4907]: I0313 14:58:01.480476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerStarted","Data":"f047cdaa087410c849606e22ef215db03f44d03ec14d4cb7f91672c24bbbade7"} Mar 13 14:58:01 crc kubenswrapper[4907]: I0313 14:58:01.786517 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:01 crc kubenswrapper[4907]: E0313 14:58:01.786730 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:02 crc kubenswrapper[4907]: I0313 14:58:02.490096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerStarted","Data":"770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0"} Mar 13 14:58:02 crc kubenswrapper[4907]: I0313 14:58:02.514527 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556898-55zxj" podStartSLOduration=1.240730842 podStartE2EDuration="2.51450287s" podCreationTimestamp="2026-03-13 14:58:00 +0000 UTC" firstStartedPulling="2026-03-13 14:58:00.8823956 +0000 UTC m=+3179.782183289" lastFinishedPulling="2026-03-13 14:58:02.156167588 +0000 UTC m=+3181.055955317" observedRunningTime="2026-03-13 14:58:02.511556289 +0000 UTC m=+3181.411343978" watchObservedRunningTime="2026-03-13 14:58:02.51450287 +0000 UTC m=+3181.414290569" Mar 13 14:58:03 crc kubenswrapper[4907]: I0313 14:58:03.503181 4907 generic.go:334] "Generic (PLEG): container finished" podID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerID="770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0" exitCode=0 Mar 13 14:58:03 crc kubenswrapper[4907]: I0313 14:58:03.503325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerDied","Data":"770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0"} Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.767370 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.861225 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.867842 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") pod \"4efb5dca-e973-443a-bcc2-3b8799fb680c\" (UID: \"4efb5dca-e973-443a-bcc2-3b8799fb680c\") " Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.873949 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556892-gq5qv"] Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.879314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf" (OuterVolumeSpecName: "kube-api-access-rjqcf") pod "4efb5dca-e973-443a-bcc2-3b8799fb680c" (UID: "4efb5dca-e973-443a-bcc2-3b8799fb680c"). InnerVolumeSpecName "kube-api-access-rjqcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 14:58:04 crc kubenswrapper[4907]: I0313 14:58:04.970760 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjqcf\" (UniqueName: \"kubernetes.io/projected/4efb5dca-e973-443a-bcc2-3b8799fb680c-kube-api-access-rjqcf\") on node \"crc\" DevicePath \"\"" Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.521429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556898-55zxj" event={"ID":"4efb5dca-e973-443a-bcc2-3b8799fb680c","Type":"ContainerDied","Data":"f047cdaa087410c849606e22ef215db03f44d03ec14d4cb7f91672c24bbbade7"} Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.521911 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f047cdaa087410c849606e22ef215db03f44d03ec14d4cb7f91672c24bbbade7" Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.521549 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556898-55zxj" Mar 13 14:58:05 crc kubenswrapper[4907]: I0313 14:58:05.793990 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbb8b9ec-85bd-4617-952d-57ce0bdb4553" path="/var/lib/kubelet/pods/cbb8b9ec-85bd-4617-952d-57ce0bdb4553/volumes" Mar 13 14:58:13 crc kubenswrapper[4907]: I0313 14:58:13.782830 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:13 crc kubenswrapper[4907]: E0313 14:58:13.783416 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:25 crc kubenswrapper[4907]: I0313 14:58:25.798765 4907 scope.go:117] "RemoveContainer" containerID="2c73e4b83765c168348350266e6ee7d5565c27f33d1acbef663a50d336ac5e75" Mar 13 14:58:26 crc kubenswrapper[4907]: I0313 14:58:26.786407 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:26 crc kubenswrapper[4907]: E0313 14:58:26.786678 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:40 crc kubenswrapper[4907]: I0313 14:58:40.782938 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:40 crc kubenswrapper[4907]: E0313 14:58:40.783705 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:58:52 crc kubenswrapper[4907]: I0313 14:58:52.782994 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:58:52 crc kubenswrapper[4907]: E0313 14:58:52.784946 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:07 crc kubenswrapper[4907]: I0313 14:59:07.783352 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:07 crc kubenswrapper[4907]: E0313 14:59:07.784162 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:19 crc kubenswrapper[4907]: I0313 14:59:19.782731 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:19 crc kubenswrapper[4907]: E0313 14:59:19.783405 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:33 crc kubenswrapper[4907]: I0313 14:59:33.782412 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:33 crc kubenswrapper[4907]: E0313 14:59:33.783164 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 14:59:47 crc kubenswrapper[4907]: I0313 14:59:47.782328 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 14:59:47 crc kubenswrapper[4907]: E0313 14:59:47.784597 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.185013 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:00:00 crc kubenswrapper[4907]: E0313 15:00:00.185744 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerName="oc" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.185761 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerName="oc" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.185969 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" containerName="oc" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.186670 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.190895 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.191259 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.192132 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.192369 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.193343 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.194470 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.202030 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.204309 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.210925 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.282874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.282990 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"auto-csr-approver-29556900-thmpg\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.283040 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.283072 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384342 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"auto-csr-approver-29556900-thmpg\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384404 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.384502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.386037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.396666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.400860 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"auto-csr-approver-29556900-thmpg\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.400923 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"collect-profiles-29556900-cxbzm\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.517742 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.523676 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:00 crc kubenswrapper[4907]: I0313 15:00:00.941768 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.005177 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:00:01 crc kubenswrapper[4907]: W0313 15:00:01.012117 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48ae7008_a336_47d8_a3da_fd198db91cba.slice/crio-f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8 WatchSource:0}: Error finding container f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8: Status 404 returned error can't find the container with id f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8 Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.301681 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556900-thmpg" event={"ID":"aafe442b-c955-456c-ac6d-a31322a5a27e","Type":"ContainerStarted","Data":"500cd8b2f75b1b643859a926fad44ef0385176e754d864d21bafc545c9b9ead7"} Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.304012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerStarted","Data":"189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431"} Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.304056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerStarted","Data":"f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8"} Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.324971 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" podStartSLOduration=1.324953346 podStartE2EDuration="1.324953346s" podCreationTimestamp="2026-03-13 15:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:00:01.32002831 +0000 UTC m=+3300.219816019" watchObservedRunningTime="2026-03-13 15:00:01.324953346 +0000 UTC m=+3300.224741035" Mar 13 15:00:01 crc kubenswrapper[4907]: I0313 15:00:01.785655 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:00:01 crc kubenswrapper[4907]: E0313 15:00:01.786264 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:00:02 crc kubenswrapper[4907]: I0313 15:00:02.311156 4907 generic.go:334] "Generic (PLEG): container finished" podID="48ae7008-a336-47d8-a3da-fd198db91cba" containerID="189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431" exitCode=0 Mar 13 15:00:02 crc kubenswrapper[4907]: I0313 15:00:02.311224 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerDied","Data":"189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431"} Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.602114 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.763422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") pod \"48ae7008-a336-47d8-a3da-fd198db91cba\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.763553 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") pod \"48ae7008-a336-47d8-a3da-fd198db91cba\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.763587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") pod \"48ae7008-a336-47d8-a3da-fd198db91cba\" (UID: \"48ae7008-a336-47d8-a3da-fd198db91cba\") " Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.764625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume" (OuterVolumeSpecName: "config-volume") pod "48ae7008-a336-47d8-a3da-fd198db91cba" (UID: "48ae7008-a336-47d8-a3da-fd198db91cba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.768930 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "48ae7008-a336-47d8-a3da-fd198db91cba" (UID: "48ae7008-a336-47d8-a3da-fd198db91cba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.769310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9" (OuterVolumeSpecName: "kube-api-access-92pc9") pod "48ae7008-a336-47d8-a3da-fd198db91cba" (UID: "48ae7008-a336-47d8-a3da-fd198db91cba"). InnerVolumeSpecName "kube-api-access-92pc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.865537 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92pc9\" (UniqueName: \"kubernetes.io/projected/48ae7008-a336-47d8-a3da-fd198db91cba-kube-api-access-92pc9\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.865663 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/48ae7008-a336-47d8-a3da-fd198db91cba-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:03 crc kubenswrapper[4907]: I0313 15:00:03.865698 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/48ae7008-a336-47d8-a3da-fd198db91cba-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.323116 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" event={"ID":"48ae7008-a336-47d8-a3da-fd198db91cba","Type":"ContainerDied","Data":"f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8"} Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.323159 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f74ce56bde217c7c03564c264b33895175d6afdd3867b429f1f77de4792d29c8" Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.323209 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm" Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.393777 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 15:00:04 crc kubenswrapper[4907]: I0313 15:00:04.398802 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556855-t8w6z"] Mar 13 15:00:05 crc kubenswrapper[4907]: I0313 15:00:05.333207 4907 generic.go:334] "Generic (PLEG): container finished" podID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerID="5fa3c1b84916d90898ac1395ff1cb1a7f88260f563edf9c1fdc38cde419411f5" exitCode=0 Mar 13 15:00:05 crc kubenswrapper[4907]: I0313 15:00:05.333299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556900-thmpg" event={"ID":"aafe442b-c955-456c-ac6d-a31322a5a27e","Type":"ContainerDied","Data":"5fa3c1b84916d90898ac1395ff1cb1a7f88260f563edf9c1fdc38cde419411f5"} Mar 13 15:00:05 crc kubenswrapper[4907]: I0313 15:00:05.791110 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80282185-9b05-4101-b906-e069f57e9bae" path="/var/lib/kubelet/pods/80282185-9b05-4101-b906-e069f57e9bae/volumes" Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.588539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.706872 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") pod \"aafe442b-c955-456c-ac6d-a31322a5a27e\" (UID: \"aafe442b-c955-456c-ac6d-a31322a5a27e\") " Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.713035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp" (OuterVolumeSpecName: "kube-api-access-fktlp") pod "aafe442b-c955-456c-ac6d-a31322a5a27e" (UID: "aafe442b-c955-456c-ac6d-a31322a5a27e"). InnerVolumeSpecName "kube-api-access-fktlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:00:06 crc kubenswrapper[4907]: I0313 15:00:06.809121 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fktlp\" (UniqueName: \"kubernetes.io/projected/aafe442b-c955-456c-ac6d-a31322a5a27e-kube-api-access-fktlp\") on node \"crc\" DevicePath \"\"" Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.350739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556900-thmpg" event={"ID":"aafe442b-c955-456c-ac6d-a31322a5a27e","Type":"ContainerDied","Data":"500cd8b2f75b1b643859a926fad44ef0385176e754d864d21bafc545c9b9ead7"} Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.351066 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="500cd8b2f75b1b643859a926fad44ef0385176e754d864d21bafc545c9b9ead7" Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.350818 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556900-thmpg" Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.648544 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.654564 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556894-8mk4g"] Mar 13 15:00:07 crc kubenswrapper[4907]: I0313 15:00:07.794844 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14af3ced-cea4-46f4-82b2-bdcc55744a4f" path="/var/lib/kubelet/pods/14af3ced-cea4-46f4-82b2-bdcc55744a4f/volumes" Mar 13 15:00:13 crc kubenswrapper[4907]: I0313 15:00:13.783613 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:00:13 crc kubenswrapper[4907]: E0313 15:00:13.784398 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:00:24 crc kubenswrapper[4907]: I0313 15:00:24.783251 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:00:25 crc kubenswrapper[4907]: I0313 15:00:25.488096 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391"} Mar 13 15:00:25 crc kubenswrapper[4907]: I0313 15:00:25.872785 4907 scope.go:117] "RemoveContainer" containerID="1113c32af3f3a8484a7765dfa4c4a4904cf0b0cad6e1387975333f03729886ff" Mar 13 15:00:25 crc kubenswrapper[4907]: I0313 15:00:25.901168 4907 scope.go:117] "RemoveContainer" containerID="1f3fec4f332e7900eb7076ad444f074ae92511e972577dd77d05eebc0931b626" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.385426 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:00:49 crc kubenswrapper[4907]: E0313 15:00:49.386293 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerName="oc" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386305 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerName="oc" Mar 13 15:00:49 crc kubenswrapper[4907]: E0313 15:00:49.386328 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" containerName="collect-profiles" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386335 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" containerName="collect-profiles" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386529 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" containerName="collect-profiles" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.386546 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" containerName="oc" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.387643 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.396199 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.542504 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.542571 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.542658 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643492 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.643859 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.644024 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.663783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"redhat-marketplace-m2b5z\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:49 crc kubenswrapper[4907]: I0313 15:00:49.707908 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.125960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.689926 4907 generic.go:334] "Generic (PLEG): container finished" podID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" exitCode=0 Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.690045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e"} Mar 13 15:00:50 crc kubenswrapper[4907]: I0313 15:00:50.690332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerStarted","Data":"6430f9fc5bca0ed430ebc1bfa36b769f8542a5c922626b992fd79be104079725"} Mar 13 15:00:52 crc kubenswrapper[4907]: I0313 15:00:52.706223 4907 generic.go:334] "Generic (PLEG): container finished" podID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" exitCode=0 Mar 13 15:00:52 crc kubenswrapper[4907]: I0313 15:00:52.706341 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b"} Mar 13 15:00:53 crc kubenswrapper[4907]: I0313 15:00:53.715837 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerStarted","Data":"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5"} Mar 13 15:00:53 crc kubenswrapper[4907]: I0313 15:00:53.736936 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m2b5z" podStartSLOduration=2.302601292 podStartE2EDuration="4.736919185s" podCreationTimestamp="2026-03-13 15:00:49 +0000 UTC" firstStartedPulling="2026-03-13 15:00:50.692777377 +0000 UTC m=+3349.592565066" lastFinishedPulling="2026-03-13 15:00:53.12709527 +0000 UTC m=+3352.026882959" observedRunningTime="2026-03-13 15:00:53.731445845 +0000 UTC m=+3352.631233554" watchObservedRunningTime="2026-03-13 15:00:53.736919185 +0000 UTC m=+3352.636706884" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.707997 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.708291 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.753764 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:00:59 crc kubenswrapper[4907]: I0313 15:00:59.839336 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:01:00 crc kubenswrapper[4907]: I0313 15:01:00.023862 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:01:01 crc kubenswrapper[4907]: I0313 15:01:01.790654 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m2b5z" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" containerID="cri-o://1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" gracePeriod=2 Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.194284 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.320838 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") pod \"98b7807e-6fa0-4bde-8f09-8289ed316462\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.321334 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") pod \"98b7807e-6fa0-4bde-8f09-8289ed316462\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.321459 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") pod \"98b7807e-6fa0-4bde-8f09-8289ed316462\" (UID: \"98b7807e-6fa0-4bde-8f09-8289ed316462\") " Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.322359 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities" (OuterVolumeSpecName: "utilities") pod "98b7807e-6fa0-4bde-8f09-8289ed316462" (UID: "98b7807e-6fa0-4bde-8f09-8289ed316462"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.328095 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4" (OuterVolumeSpecName: "kube-api-access-n9wr4") pod "98b7807e-6fa0-4bde-8f09-8289ed316462" (UID: "98b7807e-6fa0-4bde-8f09-8289ed316462"). InnerVolumeSpecName "kube-api-access-n9wr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.347634 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98b7807e-6fa0-4bde-8f09-8289ed316462" (UID: "98b7807e-6fa0-4bde-8f09-8289ed316462"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.422947 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9wr4\" (UniqueName: \"kubernetes.io/projected/98b7807e-6fa0-4bde-8f09-8289ed316462-kube-api-access-n9wr4\") on node \"crc\" DevicePath \"\"" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.422984 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.422993 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98b7807e-6fa0-4bde-8f09-8289ed316462-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797795 4907 generic.go:334] "Generic (PLEG): container finished" podID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" exitCode=0 Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797835 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5"} Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797856 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m2b5z" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797874 4907 scope.go:117] "RemoveContainer" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.797863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m2b5z" event={"ID":"98b7807e-6fa0-4bde-8f09-8289ed316462","Type":"ContainerDied","Data":"6430f9fc5bca0ed430ebc1bfa36b769f8542a5c922626b992fd79be104079725"} Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.815650 4907 scope.go:117] "RemoveContainer" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.833845 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.839170 4907 scope.go:117] "RemoveContainer" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.839558 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m2b5z"] Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.861442 4907 scope.go:117] "RemoveContainer" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" Mar 13 15:01:02 crc kubenswrapper[4907]: E0313 15:01:02.861793 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5\": container with ID starting with 1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5 not found: ID does not exist" containerID="1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.861830 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5"} err="failed to get container status \"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5\": rpc error: code = NotFound desc = could not find container \"1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5\": container with ID starting with 1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5 not found: ID does not exist" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.861860 4907 scope.go:117] "RemoveContainer" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" Mar 13 15:01:02 crc kubenswrapper[4907]: E0313 15:01:02.862379 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b\": container with ID starting with 8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b not found: ID does not exist" containerID="8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.862448 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b"} err="failed to get container status \"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b\": rpc error: code = NotFound desc = could not find container \"8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b\": container with ID starting with 8087bdb8f2e06dc3cb2d038da4ecf2ebe93fd908793fddc37698e9e8f496792b not found: ID does not exist" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.862471 4907 scope.go:117] "RemoveContainer" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" Mar 13 15:01:02 crc kubenswrapper[4907]: E0313 15:01:02.862726 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e\": container with ID starting with f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e not found: ID does not exist" containerID="f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e" Mar 13 15:01:02 crc kubenswrapper[4907]: I0313 15:01:02.862749 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e"} err="failed to get container status \"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e\": rpc error: code = NotFound desc = could not find container \"f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e\": container with ID starting with f6c782f8c5b76bea23f8182eb1e7f3a7d38bf37fbaa586bf1d7d22a54504041e not found: ID does not exist" Mar 13 15:01:03 crc kubenswrapper[4907]: I0313 15:01:03.793842 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" path="/var/lib/kubelet/pods/98b7807e-6fa0-4bde-8f09-8289ed316462/volumes" Mar 13 15:01:10 crc kubenswrapper[4907]: E0313 15:01:10.990951 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:21 crc kubenswrapper[4907]: E0313 15:01:21.175499 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:31 crc kubenswrapper[4907]: E0313 15:01:31.332831 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:41 crc kubenswrapper[4907]: E0313 15:01:41.483739 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:01:51 crc kubenswrapper[4907]: E0313 15:01:51.670101 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.139651 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:02:00 crc kubenswrapper[4907]: E0313 15:02:00.142065 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-content" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142165 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-content" Mar 13 15:02:00 crc kubenswrapper[4907]: E0313 15:02:00.142258 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142317 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" Mar 13 15:02:00 crc kubenswrapper[4907]: E0313 15:02:00.142483 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-utilities" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142560 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="extract-utilities" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.142745 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="98b7807e-6fa0-4bde-8f09-8289ed316462" containerName="registry-server" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.143377 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.146928 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.146986 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.148620 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.149054 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.257325 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"auto-csr-approver-29556902-92pzz\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.358634 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"auto-csr-approver-29556902-92pzz\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.376121 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"auto-csr-approver-29556902-92pzz\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.469351 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.924397 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:02:00 crc kubenswrapper[4907]: I0313 15:02:00.935203 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:02:01 crc kubenswrapper[4907]: I0313 15:02:01.257369 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerStarted","Data":"bd9a11fb2fb4bd3d8600ccfeb75929b4157aca3412cb52bb36e9d792add6c662"} Mar 13 15:02:01 crc kubenswrapper[4907]: E0313 15:02:01.853672 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-conmon-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98b7807e_6fa0_4bde_8f09_8289ed316462.slice/crio-1b9d8269123996da45f8c0e1ef845fe700956a76fbab44efafea4a51c69437b5.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:02:02 crc kubenswrapper[4907]: I0313 15:02:02.268777 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerStarted","Data":"be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b"} Mar 13 15:02:02 crc kubenswrapper[4907]: I0313 15:02:02.284871 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556902-92pzz" podStartSLOduration=1.3925067819999999 podStartE2EDuration="2.284841431s" podCreationTimestamp="2026-03-13 15:02:00 +0000 UTC" firstStartedPulling="2026-03-13 15:02:00.934531009 +0000 UTC m=+3419.834318688" lastFinishedPulling="2026-03-13 15:02:01.826865648 +0000 UTC m=+3420.726653337" observedRunningTime="2026-03-13 15:02:02.280559423 +0000 UTC m=+3421.180347112" watchObservedRunningTime="2026-03-13 15:02:02.284841431 +0000 UTC m=+3421.184629120" Mar 13 15:02:03 crc kubenswrapper[4907]: I0313 15:02:03.278676 4907 generic.go:334] "Generic (PLEG): container finished" podID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerID="be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b" exitCode=0 Mar 13 15:02:03 crc kubenswrapper[4907]: I0313 15:02:03.278771 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerDied","Data":"be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b"} Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.586305 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.724658 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") pod \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\" (UID: \"076c2b87-8e1f-4dba-9b60-eb24660aa1f1\") " Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.730294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h" (OuterVolumeSpecName: "kube-api-access-qxf2h") pod "076c2b87-8e1f-4dba-9b60-eb24660aa1f1" (UID: "076c2b87-8e1f-4dba-9b60-eb24660aa1f1"). InnerVolumeSpecName "kube-api-access-qxf2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.830654 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxf2h\" (UniqueName: \"kubernetes.io/projected/076c2b87-8e1f-4dba-9b60-eb24660aa1f1-kube-api-access-qxf2h\") on node \"crc\" DevicePath \"\"" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.872573 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:04.880988 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556896-7dlx2"] Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.292605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556902-92pzz" event={"ID":"076c2b87-8e1f-4dba-9b60-eb24660aa1f1","Type":"ContainerDied","Data":"bd9a11fb2fb4bd3d8600ccfeb75929b4157aca3412cb52bb36e9d792add6c662"} Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.292669 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd9a11fb2fb4bd3d8600ccfeb75929b4157aca3412cb52bb36e9d792add6c662" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.292669 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556902-92pzz" Mar 13 15:02:05 crc kubenswrapper[4907]: I0313 15:02:05.790448 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e193f9f5-a033-4f42-9932-62f7b4edd0b2" path="/var/lib/kubelet/pods/e193f9f5-a033-4f42-9932-62f7b4edd0b2/volumes" Mar 13 15:02:25 crc kubenswrapper[4907]: I0313 15:02:25.996314 4907 scope.go:117] "RemoveContainer" containerID="19c525d3167fb54f80c181f5117cf495aa5c0602e81951a3395f3a05b6fbc6c9" Mar 13 15:02:48 crc kubenswrapper[4907]: I0313 15:02:48.041668 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:02:48 crc kubenswrapper[4907]: I0313 15:02:48.042261 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:03:18 crc kubenswrapper[4907]: I0313 15:03:18.042455 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:03:18 crc kubenswrapper[4907]: I0313 15:03:18.043077 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.772415 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:27 crc kubenswrapper[4907]: E0313 15:03:27.774189 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerName="oc" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.774275 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerName="oc" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.774550 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" containerName="oc" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.775977 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.800243 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.887784 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.888359 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.888512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.989513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.989565 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.989618 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.990164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:27 crc kubenswrapper[4907]: I0313 15:03:27.990279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.011233 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"redhat-operators-66vvh\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.123541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.548091 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.910878 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerID="9b879515804206ce1b2b94caf0c492d6383320b9348237ccf02d879b6d4dc8ff" exitCode=0 Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.910968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"9b879515804206ce1b2b94caf0c492d6383320b9348237ccf02d879b6d4dc8ff"} Mar 13 15:03:28 crc kubenswrapper[4907]: I0313 15:03:28.911126 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerStarted","Data":"ded7a4a7c9ba3e3a3c449453acd07df91c27c365325bee0b5ef808e168610d02"} Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.172465 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.174842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.192040 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.318208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.318262 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.318866 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420156 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420222 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420301 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420796 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.420896 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.440905 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"certified-operators-gghnk\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.493688 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.926577 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerID="0ebf89d94f759be904e95ad5288965a4b932133de012c3259bfbd8638ad7acdb" exitCode=0 Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.926683 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"0ebf89d94f759be904e95ad5288965a4b932133de012c3259bfbd8638ad7acdb"} Mar 13 15:03:30 crc kubenswrapper[4907]: W0313 15:03:30.982138 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58109682_9dbe_4b9f_ac68_436adf1e5943.slice/crio-a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05 WatchSource:0}: Error finding container a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05: Status 404 returned error can't find the container with id a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05 Mar 13 15:03:30 crc kubenswrapper[4907]: I0313 15:03:30.986459 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.935029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerStarted","Data":"554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb"} Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.936237 4907 generic.go:334] "Generic (PLEG): container finished" podID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" exitCode=0 Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.936268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e"} Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.936285 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerStarted","Data":"a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05"} Mar 13 15:03:31 crc kubenswrapper[4907]: I0313 15:03:31.952556 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-66vvh" podStartSLOduration=2.34704898 podStartE2EDuration="4.952532762s" podCreationTimestamp="2026-03-13 15:03:27 +0000 UTC" firstStartedPulling="2026-03-13 15:03:28.912746037 +0000 UTC m=+3507.812533726" lastFinishedPulling="2026-03-13 15:03:31.518229819 +0000 UTC m=+3510.418017508" observedRunningTime="2026-03-13 15:03:31.949317465 +0000 UTC m=+3510.849105154" watchObservedRunningTime="2026-03-13 15:03:31.952532762 +0000 UTC m=+3510.852320451" Mar 13 15:03:37 crc kubenswrapper[4907]: I0313 15:03:37.973612 4907 generic.go:334] "Generic (PLEG): container finished" podID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" exitCode=0 Mar 13 15:03:37 crc kubenswrapper[4907]: I0313 15:03:37.973674 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902"} Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.124394 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.124448 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.161426 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:38 crc kubenswrapper[4907]: I0313 15:03:38.983236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerStarted","Data":"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310"} Mar 13 15:03:39 crc kubenswrapper[4907]: I0313 15:03:39.004832 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gghnk" podStartSLOduration=2.481883791 podStartE2EDuration="9.004811144s" podCreationTimestamp="2026-03-13 15:03:30 +0000 UTC" firstStartedPulling="2026-03-13 15:03:31.938010715 +0000 UTC m=+3510.837798404" lastFinishedPulling="2026-03-13 15:03:38.460938068 +0000 UTC m=+3517.360725757" observedRunningTime="2026-03-13 15:03:39.004811834 +0000 UTC m=+3517.904599563" watchObservedRunningTime="2026-03-13 15:03:39.004811144 +0000 UTC m=+3517.904598833" Mar 13 15:03:39 crc kubenswrapper[4907]: I0313 15:03:39.028545 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.210021 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.494924 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.494973 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.579409 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:40 crc kubenswrapper[4907]: I0313 15:03:40.995621 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-66vvh" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" containerID="cri-o://554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb" gracePeriod=2 Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.004246 4907 generic.go:334] "Generic (PLEG): container finished" podID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerID="554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb" exitCode=0 Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.004327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb"} Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.477267 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.589683 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") pod \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.589861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") pod \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.589931 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") pod \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\" (UID: \"e7a96a53-ed40-4b76-b1f3-58edd7dec59c\") " Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.591174 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities" (OuterVolumeSpecName: "utilities") pod "e7a96a53-ed40-4b76-b1f3-58edd7dec59c" (UID: "e7a96a53-ed40-4b76-b1f3-58edd7dec59c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.591477 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.596967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg" (OuterVolumeSpecName: "kube-api-access-j5wrg") pod "e7a96a53-ed40-4b76-b1f3-58edd7dec59c" (UID: "e7a96a53-ed40-4b76-b1f3-58edd7dec59c"). InnerVolumeSpecName "kube-api-access-j5wrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.702488 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5wrg\" (UniqueName: \"kubernetes.io/projected/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-kube-api-access-j5wrg\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.744175 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7a96a53-ed40-4b76-b1f3-58edd7dec59c" (UID: "e7a96a53-ed40-4b76-b1f3-58edd7dec59c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:42 crc kubenswrapper[4907]: I0313 15:03:42.804025 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7a96a53-ed40-4b76-b1f3-58edd7dec59c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.016009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-66vvh" event={"ID":"e7a96a53-ed40-4b76-b1f3-58edd7dec59c","Type":"ContainerDied","Data":"ded7a4a7c9ba3e3a3c449453acd07df91c27c365325bee0b5ef808e168610d02"} Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.016063 4907 scope.go:117] "RemoveContainer" containerID="554246066c0b4d2e5bcab179363cf1a580411eeb84336e62f65f4c2957f75cfb" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.016219 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-66vvh" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.049453 4907 scope.go:117] "RemoveContainer" containerID="0ebf89d94f759be904e95ad5288965a4b932133de012c3259bfbd8638ad7acdb" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.052787 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.058063 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-66vvh"] Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.079567 4907 scope.go:117] "RemoveContainer" containerID="9b879515804206ce1b2b94caf0c492d6383320b9348237ccf02d879b6d4dc8ff" Mar 13 15:03:43 crc kubenswrapper[4907]: I0313 15:03:43.791397 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" path="/var/lib/kubelet/pods/e7a96a53-ed40-4b76-b1f3-58edd7dec59c/volumes" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.041948 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.042540 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.042590 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.043201 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:03:48 crc kubenswrapper[4907]: I0313 15:03:48.043258 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391" gracePeriod=600 Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.062815 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391" exitCode=0 Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.062920 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391"} Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.063186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a"} Mar 13 15:03:49 crc kubenswrapper[4907]: I0313 15:03:49.063229 4907 scope.go:117] "RemoveContainer" containerID="a44259e1e0a7edc8acdd625ba3f7415d70aa78b760d8eeb81254942722d3acb3" Mar 13 15:03:50 crc kubenswrapper[4907]: I0313 15:03:50.539109 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:50 crc kubenswrapper[4907]: I0313 15:03:50.587400 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.082074 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gghnk" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" containerID="cri-o://24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" gracePeriod=2 Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.476302 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.631397 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") pod \"58109682-9dbe-4b9f-ac68-436adf1e5943\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.631479 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") pod \"58109682-9dbe-4b9f-ac68-436adf1e5943\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.631551 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") pod \"58109682-9dbe-4b9f-ac68-436adf1e5943\" (UID: \"58109682-9dbe-4b9f-ac68-436adf1e5943\") " Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.632516 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities" (OuterVolumeSpecName: "utilities") pod "58109682-9dbe-4b9f-ac68-436adf1e5943" (UID: "58109682-9dbe-4b9f-ac68-436adf1e5943"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.642065 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g" (OuterVolumeSpecName: "kube-api-access-4m69g") pod "58109682-9dbe-4b9f-ac68-436adf1e5943" (UID: "58109682-9dbe-4b9f-ac68-436adf1e5943"). InnerVolumeSpecName "kube-api-access-4m69g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.690500 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58109682-9dbe-4b9f-ac68-436adf1e5943" (UID: "58109682-9dbe-4b9f-ac68-436adf1e5943"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.733712 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.733762 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m69g\" (UniqueName: \"kubernetes.io/projected/58109682-9dbe-4b9f-ac68-436adf1e5943-kube-api-access-4m69g\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:51 crc kubenswrapper[4907]: I0313 15:03:51.733781 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58109682-9dbe-4b9f-ac68-436adf1e5943-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091216 4907 generic.go:334] "Generic (PLEG): container finished" podID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" exitCode=0 Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310"} Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091310 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gghnk" event={"ID":"58109682-9dbe-4b9f-ac68-436adf1e5943","Type":"ContainerDied","Data":"a6de591612bd7f598975b473588315c1b21aeee80aebf09d8fc818c82ba74d05"} Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091315 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gghnk" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.091336 4907 scope.go:117] "RemoveContainer" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.112965 4907 scope.go:117] "RemoveContainer" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.114907 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.122439 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gghnk"] Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.146786 4907 scope.go:117] "RemoveContainer" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.167188 4907 scope.go:117] "RemoveContainer" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" Mar 13 15:03:52 crc kubenswrapper[4907]: E0313 15:03:52.167588 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310\": container with ID starting with 24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310 not found: ID does not exist" containerID="24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.167723 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310"} err="failed to get container status \"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310\": rpc error: code = NotFound desc = could not find container \"24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310\": container with ID starting with 24c5f8f053b603164ac21c58d2935e095541de88fc62f30e30f087b6e3342310 not found: ID does not exist" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.167799 4907 scope.go:117] "RemoveContainer" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" Mar 13 15:03:52 crc kubenswrapper[4907]: E0313 15:03:52.168303 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902\": container with ID starting with b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902 not found: ID does not exist" containerID="b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.168384 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902"} err="failed to get container status \"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902\": rpc error: code = NotFound desc = could not find container \"b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902\": container with ID starting with b74e9e26c2ace1b371fd700c4c00135c61ddd6fae9f88bedfe17d9937c863902 not found: ID does not exist" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.168446 4907 scope.go:117] "RemoveContainer" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" Mar 13 15:03:52 crc kubenswrapper[4907]: E0313 15:03:52.168804 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e\": container with ID starting with b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e not found: ID does not exist" containerID="b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e" Mar 13 15:03:52 crc kubenswrapper[4907]: I0313 15:03:52.168914 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e"} err="failed to get container status \"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e\": rpc error: code = NotFound desc = could not find container \"b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e\": container with ID starting with b15bbdd7e88ef6a86f31df0036fa2f21235ebe45494d497ad718bac5b50fb58e not found: ID does not exist" Mar 13 15:03:53 crc kubenswrapper[4907]: I0313 15:03:53.791131 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" path="/var/lib/kubelet/pods/58109682-9dbe-4b9f-ac68-436adf1e5943/volumes" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.383716 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385235 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385253 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385271 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385278 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385286 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385292 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="extract-content" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385301 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385307 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="extract-utilities" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385324 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385329 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: E0313 15:03:55.385345 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385351 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385492 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7a96a53-ed40-4b76-b1f3-58edd7dec59c" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.385510 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="58109682-9dbe-4b9f-ac68-436adf1e5943" containerName="registry-server" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.386497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.438743 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.495290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.495382 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.495432 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.596503 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.596588 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.596642 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.597127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.597179 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.624106 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"community-operators-zx5wb\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:55 crc kubenswrapper[4907]: I0313 15:03:55.710858 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:03:56 crc kubenswrapper[4907]: I0313 15:03:56.016274 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:03:56 crc kubenswrapper[4907]: I0313 15:03:56.118973 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerStarted","Data":"0bc7f96bf4dee2dfb710654f87828dd414be19d802599fada94514556b97115a"} Mar 13 15:03:57 crc kubenswrapper[4907]: I0313 15:03:57.126293 4907 generic.go:334] "Generic (PLEG): container finished" podID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" exitCode=0 Mar 13 15:03:57 crc kubenswrapper[4907]: I0313 15:03:57.126413 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775"} Mar 13 15:03:58 crc kubenswrapper[4907]: I0313 15:03:58.135405 4907 generic.go:334] "Generic (PLEG): container finished" podID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" exitCode=0 Mar 13 15:03:58 crc kubenswrapper[4907]: I0313 15:03:58.135466 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5"} Mar 13 15:03:59 crc kubenswrapper[4907]: I0313 15:03:59.144849 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerStarted","Data":"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88"} Mar 13 15:03:59 crc kubenswrapper[4907]: I0313 15:03:59.170625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zx5wb" podStartSLOduration=2.5354631789999997 podStartE2EDuration="4.170608146s" podCreationTimestamp="2026-03-13 15:03:55 +0000 UTC" firstStartedPulling="2026-03-13 15:03:57.128076673 +0000 UTC m=+3536.027864362" lastFinishedPulling="2026-03-13 15:03:58.76322164 +0000 UTC m=+3537.663009329" observedRunningTime="2026-03-13 15:03:59.16054434 +0000 UTC m=+3538.060332029" watchObservedRunningTime="2026-03-13 15:03:59.170608146 +0000 UTC m=+3538.070395835" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.139036 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.140736 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.147762 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.148838 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.153044 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.167349 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.270119 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"auto-csr-approver-29556904-4qqhq\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.372150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"auto-csr-approver-29556904-4qqhq\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.401845 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"auto-csr-approver-29556904-4qqhq\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.474539 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:00 crc kubenswrapper[4907]: I0313 15:04:00.900061 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:04:00 crc kubenswrapper[4907]: W0313 15:04:00.906279 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod856cd90a_f3c2_4b46_b5eb_77e85004dbf6.slice/crio-51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe WatchSource:0}: Error finding container 51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe: Status 404 returned error can't find the container with id 51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe Mar 13 15:04:01 crc kubenswrapper[4907]: I0313 15:04:01.158432 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerStarted","Data":"51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe"} Mar 13 15:04:02 crc kubenswrapper[4907]: I0313 15:04:02.167342 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerStarted","Data":"6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594"} Mar 13 15:04:02 crc kubenswrapper[4907]: I0313 15:04:02.191616 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" podStartSLOduration=1.378680985 podStartE2EDuration="2.191587665s" podCreationTimestamp="2026-03-13 15:04:00 +0000 UTC" firstStartedPulling="2026-03-13 15:04:00.90820107 +0000 UTC m=+3539.807988759" lastFinishedPulling="2026-03-13 15:04:01.72110772 +0000 UTC m=+3540.620895439" observedRunningTime="2026-03-13 15:04:02.189366795 +0000 UTC m=+3541.089154494" watchObservedRunningTime="2026-03-13 15:04:02.191587665 +0000 UTC m=+3541.091375384" Mar 13 15:04:03 crc kubenswrapper[4907]: I0313 15:04:03.175593 4907 generic.go:334] "Generic (PLEG): container finished" podID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerID="6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594" exitCode=0 Mar 13 15:04:03 crc kubenswrapper[4907]: I0313 15:04:03.175651 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerDied","Data":"6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594"} Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.439809 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.628922 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") pod \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\" (UID: \"856cd90a-f3c2-4b46-b5eb-77e85004dbf6\") " Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.634290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5" (OuterVolumeSpecName: "kube-api-access-c67s5") pod "856cd90a-f3c2-4b46-b5eb-77e85004dbf6" (UID: "856cd90a-f3c2-4b46-b5eb-77e85004dbf6"). InnerVolumeSpecName "kube-api-access-c67s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.730475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c67s5\" (UniqueName: \"kubernetes.io/projected/856cd90a-f3c2-4b46-b5eb-77e85004dbf6-kube-api-access-c67s5\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.869366 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 15:04:04 crc kubenswrapper[4907]: I0313 15:04:04.875031 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556898-55zxj"] Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.196724 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" event={"ID":"856cd90a-f3c2-4b46-b5eb-77e85004dbf6","Type":"ContainerDied","Data":"51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe"} Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.196766 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556904-4qqhq" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.196773 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d2b9d00ee0e8fc8c6c0d1f2930961379e1f2ed812db0fb202c46d98f7acbbe" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.711921 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.712190 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.757609 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:05 crc kubenswrapper[4907]: I0313 15:04:05.790737 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4efb5dca-e973-443a-bcc2-3b8799fb680c" path="/var/lib/kubelet/pods/4efb5dca-e973-443a-bcc2-3b8799fb680c/volumes" Mar 13 15:04:06 crc kubenswrapper[4907]: I0313 15:04:06.254038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:06 crc kubenswrapper[4907]: I0313 15:04:06.302658 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:04:08 crc kubenswrapper[4907]: I0313 15:04:08.215613 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zx5wb" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" containerID="cri-o://83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" gracePeriod=2 Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.108929 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223120 4907 generic.go:334] "Generic (PLEG): container finished" podID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" exitCode=0 Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223173 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88"} Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223184 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zx5wb" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223204 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zx5wb" event={"ID":"bdee4689-2396-4e32-ae39-6ffc60033f38","Type":"ContainerDied","Data":"0bc7f96bf4dee2dfb710654f87828dd414be19d802599fada94514556b97115a"} Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.223224 4907 scope.go:117] "RemoveContainer" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.240464 4907 scope.go:117] "RemoveContainer" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.257224 4907 scope.go:117] "RemoveContainer" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.284054 4907 scope.go:117] "RemoveContainer" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" Mar 13 15:04:09 crc kubenswrapper[4907]: E0313 15:04:09.286606 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88\": container with ID starting with 83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88 not found: ID does not exist" containerID="83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.286652 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88"} err="failed to get container status \"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88\": rpc error: code = NotFound desc = could not find container \"83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88\": container with ID starting with 83fa90112c03ad8ce15458ee0ff9f3e2e51c4749967eee63928ff69aae2e9b88 not found: ID does not exist" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.286676 4907 scope.go:117] "RemoveContainer" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" Mar 13 15:04:09 crc kubenswrapper[4907]: E0313 15:04:09.287223 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5\": container with ID starting with 84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5 not found: ID does not exist" containerID="84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.287253 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5"} err="failed to get container status \"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5\": rpc error: code = NotFound desc = could not find container \"84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5\": container with ID starting with 84f8ced411df60770107c3973266fb82290633ef4138573f621b9eb0d07085e5 not found: ID does not exist" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.287267 4907 scope.go:117] "RemoveContainer" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" Mar 13 15:04:09 crc kubenswrapper[4907]: E0313 15:04:09.288131 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775\": container with ID starting with 6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775 not found: ID does not exist" containerID="6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.288160 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775"} err="failed to get container status \"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775\": rpc error: code = NotFound desc = could not find container \"6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775\": container with ID starting with 6568d87c6bad18c68d28d286c5ed63b49b3f4a5ac95f46d9d41dc5bd03f93775 not found: ID does not exist" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.290623 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") pod \"bdee4689-2396-4e32-ae39-6ffc60033f38\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.290713 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") pod \"bdee4689-2396-4e32-ae39-6ffc60033f38\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.290826 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") pod \"bdee4689-2396-4e32-ae39-6ffc60033f38\" (UID: \"bdee4689-2396-4e32-ae39-6ffc60033f38\") " Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.291790 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities" (OuterVolumeSpecName: "utilities") pod "bdee4689-2396-4e32-ae39-6ffc60033f38" (UID: "bdee4689-2396-4e32-ae39-6ffc60033f38"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.295475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql" (OuterVolumeSpecName: "kube-api-access-q7dql") pod "bdee4689-2396-4e32-ae39-6ffc60033f38" (UID: "bdee4689-2396-4e32-ae39-6ffc60033f38"). InnerVolumeSpecName "kube-api-access-q7dql". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.345566 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bdee4689-2396-4e32-ae39-6ffc60033f38" (UID: "bdee4689-2396-4e32-ae39-6ffc60033f38"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.392169 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.392203 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7dql\" (UniqueName: \"kubernetes.io/projected/bdee4689-2396-4e32-ae39-6ffc60033f38-kube-api-access-q7dql\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.392215 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bdee4689-2396-4e32-ae39-6ffc60033f38-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.553556 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.563115 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zx5wb"] Mar 13 15:04:09 crc kubenswrapper[4907]: I0313 15:04:09.790218 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" path="/var/lib/kubelet/pods/bdee4689-2396-4e32-ae39-6ffc60033f38/volumes" Mar 13 15:04:26 crc kubenswrapper[4907]: I0313 15:04:26.075695 4907 scope.go:117] "RemoveContainer" containerID="770adaf80533ac339976a419aecf995f45ec4e4221e3c81f0b0f3f79b4736ef0" Mar 13 15:05:48 crc kubenswrapper[4907]: I0313 15:05:48.041325 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:05:48 crc kubenswrapper[4907]: I0313 15:05:48.041833 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.148927 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.151977 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-utilities" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152137 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-utilities" Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.152262 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.152495 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerName="oc" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152612 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerName="oc" Mar 13 15:06:00 crc kubenswrapper[4907]: E0313 15:06:00.152729 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-content" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.152905 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="extract-content" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.153323 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdee4689-2396-4e32-ae39-6ffc60033f38" containerName="registry-server" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.153536 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" containerName="oc" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.154409 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.157930 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.158549 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.158658 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.161239 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.309705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"auto-csr-approver-29556906-kqncj\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.411677 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"auto-csr-approver-29556906-kqncj\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.450838 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"auto-csr-approver-29556906-kqncj\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.473206 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:00 crc kubenswrapper[4907]: I0313 15:06:00.911187 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:06:01 crc kubenswrapper[4907]: I0313 15:06:01.439667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556906-kqncj" event={"ID":"7a501d2e-0774-414e-8afb-39a9a9d8b731","Type":"ContainerStarted","Data":"a11a599b13906f7f8bcbe895f454a7ae524d8105c5bce3da9b1f7d5287818788"} Mar 13 15:06:02 crc kubenswrapper[4907]: I0313 15:06:02.448973 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerID="c0f40228828ee5195fb41b365b21febab7b0feb3cf02e777e21a472c7ba50dbb" exitCode=0 Mar 13 15:06:02 crc kubenswrapper[4907]: I0313 15:06:02.449087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556906-kqncj" event={"ID":"7a501d2e-0774-414e-8afb-39a9a9d8b731","Type":"ContainerDied","Data":"c0f40228828ee5195fb41b365b21febab7b0feb3cf02e777e21a472c7ba50dbb"} Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.721515 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.859085 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") pod \"7a501d2e-0774-414e-8afb-39a9a9d8b731\" (UID: \"7a501d2e-0774-414e-8afb-39a9a9d8b731\") " Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.864415 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm" (OuterVolumeSpecName: "kube-api-access-ph8jm") pod "7a501d2e-0774-414e-8afb-39a9a9d8b731" (UID: "7a501d2e-0774-414e-8afb-39a9a9d8b731"). InnerVolumeSpecName "kube-api-access-ph8jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:06:03 crc kubenswrapper[4907]: I0313 15:06:03.960969 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph8jm\" (UniqueName: \"kubernetes.io/projected/7a501d2e-0774-414e-8afb-39a9a9d8b731-kube-api-access-ph8jm\") on node \"crc\" DevicePath \"\"" Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.463187 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556906-kqncj" event={"ID":"7a501d2e-0774-414e-8afb-39a9a9d8b731","Type":"ContainerDied","Data":"a11a599b13906f7f8bcbe895f454a7ae524d8105c5bce3da9b1f7d5287818788"} Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.463232 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a11a599b13906f7f8bcbe895f454a7ae524d8105c5bce3da9b1f7d5287818788" Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.463249 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556906-kqncj" Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.782140 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:06:04 crc kubenswrapper[4907]: I0313 15:06:04.788721 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556900-thmpg"] Mar 13 15:06:05 crc kubenswrapper[4907]: I0313 15:06:05.799943 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aafe442b-c955-456c-ac6d-a31322a5a27e" path="/var/lib/kubelet/pods/aafe442b-c955-456c-ac6d-a31322a5a27e/volumes" Mar 13 15:06:18 crc kubenswrapper[4907]: I0313 15:06:18.041600 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:06:18 crc kubenswrapper[4907]: I0313 15:06:18.042306 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:06:26 crc kubenswrapper[4907]: I0313 15:06:26.189604 4907 scope.go:117] "RemoveContainer" containerID="5fa3c1b84916d90898ac1395ff1cb1a7f88260f563edf9c1fdc38cde419411f5" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.042263 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.042794 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.042848 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.043501 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.043553 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" gracePeriod=600 Mar 13 15:06:48 crc kubenswrapper[4907]: E0313 15:06:48.164973 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.788688 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" exitCode=0 Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.788745 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a"} Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.789116 4907 scope.go:117] "RemoveContainer" containerID="62e3208576ab64ce53d2a23023d86e9e8cd79b456c2ef95c5adcb1299b258391" Mar 13 15:06:48 crc kubenswrapper[4907]: I0313 15:06:48.790980 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:06:48 crc kubenswrapper[4907]: E0313 15:06:48.791577 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:03 crc kubenswrapper[4907]: I0313 15:07:03.782013 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:03 crc kubenswrapper[4907]: E0313 15:07:03.782780 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:17 crc kubenswrapper[4907]: I0313 15:07:17.783019 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:17 crc kubenswrapper[4907]: E0313 15:07:17.783697 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:30 crc kubenswrapper[4907]: I0313 15:07:30.782645 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:30 crc kubenswrapper[4907]: E0313 15:07:30.783472 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:44 crc kubenswrapper[4907]: I0313 15:07:44.782945 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:44 crc kubenswrapper[4907]: E0313 15:07:44.783571 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:07:57 crc kubenswrapper[4907]: I0313 15:07:57.783515 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:07:57 crc kubenswrapper[4907]: E0313 15:07:57.784279 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.143746 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:08:00 crc kubenswrapper[4907]: E0313 15:08:00.144184 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerName="oc" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.144201 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerName="oc" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.144371 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" containerName="oc" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.144928 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.146993 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.147242 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.147772 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.155552 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.180710 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"auto-csr-approver-29556908-lsqcs\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.282690 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"auto-csr-approver-29556908-lsqcs\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.299972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"auto-csr-approver-29556908-lsqcs\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.468703 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.871357 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:08:00 crc kubenswrapper[4907]: I0313 15:08:00.877257 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:08:01 crc kubenswrapper[4907]: I0313 15:08:01.325595 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" event={"ID":"555f4e30-b850-428b-a0fc-e15af18a58be","Type":"ContainerStarted","Data":"a174107a87cee85d700bd1c92ce4cce88bc356e63d464a68d165ca89b478e124"} Mar 13 15:08:03 crc kubenswrapper[4907]: I0313 15:08:03.342683 4907 generic.go:334] "Generic (PLEG): container finished" podID="555f4e30-b850-428b-a0fc-e15af18a58be" containerID="b199bdf3c496349ef8f75e1af0202d9bd0ccc18b9b13016f199b8123d6045267" exitCode=0 Mar 13 15:08:03 crc kubenswrapper[4907]: I0313 15:08:03.342740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" event={"ID":"555f4e30-b850-428b-a0fc-e15af18a58be","Type":"ContainerDied","Data":"b199bdf3c496349ef8f75e1af0202d9bd0ccc18b9b13016f199b8123d6045267"} Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.610642 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.649392 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") pod \"555f4e30-b850-428b-a0fc-e15af18a58be\" (UID: \"555f4e30-b850-428b-a0fc-e15af18a58be\") " Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.654544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt" (OuterVolumeSpecName: "kube-api-access-cl9zt") pod "555f4e30-b850-428b-a0fc-e15af18a58be" (UID: "555f4e30-b850-428b-a0fc-e15af18a58be"). InnerVolumeSpecName "kube-api-access-cl9zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:08:04 crc kubenswrapper[4907]: I0313 15:08:04.750722 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl9zt\" (UniqueName: \"kubernetes.io/projected/555f4e30-b850-428b-a0fc-e15af18a58be-kube-api-access-cl9zt\") on node \"crc\" DevicePath \"\"" Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.356852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" event={"ID":"555f4e30-b850-428b-a0fc-e15af18a58be","Type":"ContainerDied","Data":"a174107a87cee85d700bd1c92ce4cce88bc356e63d464a68d165ca89b478e124"} Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.357098 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a174107a87cee85d700bd1c92ce4cce88bc356e63d464a68d165ca89b478e124" Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.356949 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556908-lsqcs" Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.675222 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.681109 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556902-92pzz"] Mar 13 15:08:05 crc kubenswrapper[4907]: I0313 15:08:05.792757 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="076c2b87-8e1f-4dba-9b60-eb24660aa1f1" path="/var/lib/kubelet/pods/076c2b87-8e1f-4dba-9b60-eb24660aa1f1/volumes" Mar 13 15:08:12 crc kubenswrapper[4907]: I0313 15:08:12.782190 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:12 crc kubenswrapper[4907]: E0313 15:08:12.782997 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:26 crc kubenswrapper[4907]: I0313 15:08:26.270916 4907 scope.go:117] "RemoveContainer" containerID="be729cf40227f8298124611ddf4a79928c79ddf05e8dab4417abc40fd127f33b" Mar 13 15:08:26 crc kubenswrapper[4907]: I0313 15:08:26.782469 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:26 crc kubenswrapper[4907]: E0313 15:08:26.783017 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:41 crc kubenswrapper[4907]: I0313 15:08:41.787201 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:41 crc kubenswrapper[4907]: E0313 15:08:41.788000 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:08:54 crc kubenswrapper[4907]: I0313 15:08:54.783203 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:08:54 crc kubenswrapper[4907]: E0313 15:08:54.784306 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:07 crc kubenswrapper[4907]: I0313 15:09:07.782047 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:07 crc kubenswrapper[4907]: E0313 15:09:07.782838 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:18 crc kubenswrapper[4907]: I0313 15:09:18.782540 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:18 crc kubenswrapper[4907]: E0313 15:09:18.783140 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:31 crc kubenswrapper[4907]: I0313 15:09:31.790525 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:31 crc kubenswrapper[4907]: E0313 15:09:31.791208 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:42 crc kubenswrapper[4907]: I0313 15:09:42.782732 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:42 crc kubenswrapper[4907]: E0313 15:09:42.783629 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:09:57 crc kubenswrapper[4907]: I0313 15:09:57.782752 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:09:57 crc kubenswrapper[4907]: E0313 15:09:57.784155 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.148097 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:10:00 crc kubenswrapper[4907]: E0313 15:10:00.148547 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" containerName="oc" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.148565 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" containerName="oc" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.148814 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" containerName="oc" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.149509 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.151380 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.152026 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.152359 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.164512 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.285597 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"auto-csr-approver-29556910-5pmwx\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.387073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"auto-csr-approver-29556910-5pmwx\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.425134 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"auto-csr-approver-29556910-5pmwx\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.468050 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:00 crc kubenswrapper[4907]: I0313 15:10:00.712852 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:10:00 crc kubenswrapper[4907]: W0313 15:10:00.717251 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode74e8820_9ca4_4010_9070_ef32010a85c8.slice/crio-db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d WatchSource:0}: Error finding container db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d: Status 404 returned error can't find the container with id db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d Mar 13 15:10:01 crc kubenswrapper[4907]: I0313 15:10:01.233503 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerStarted","Data":"db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d"} Mar 13 15:10:04 crc kubenswrapper[4907]: I0313 15:10:04.254523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerStarted","Data":"6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce"} Mar 13 15:10:04 crc kubenswrapper[4907]: I0313 15:10:04.270839 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" podStartSLOduration=1.299812045 podStartE2EDuration="4.270808645s" podCreationTimestamp="2026-03-13 15:10:00 +0000 UTC" firstStartedPulling="2026-03-13 15:10:00.720067765 +0000 UTC m=+3899.619855464" lastFinishedPulling="2026-03-13 15:10:03.691064335 +0000 UTC m=+3902.590852064" observedRunningTime="2026-03-13 15:10:04.267683409 +0000 UTC m=+3903.167471098" watchObservedRunningTime="2026-03-13 15:10:04.270808645 +0000 UTC m=+3903.170596334" Mar 13 15:10:05 crc kubenswrapper[4907]: I0313 15:10:05.263337 4907 generic.go:334] "Generic (PLEG): container finished" podID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerID="6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce" exitCode=0 Mar 13 15:10:05 crc kubenswrapper[4907]: I0313 15:10:05.263439 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerDied","Data":"6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce"} Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.536432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.682067 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") pod \"e74e8820-9ca4-4010-9070-ef32010a85c8\" (UID: \"e74e8820-9ca4-4010-9070-ef32010a85c8\") " Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.688296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l" (OuterVolumeSpecName: "kube-api-access-8kv6l") pod "e74e8820-9ca4-4010-9070-ef32010a85c8" (UID: "e74e8820-9ca4-4010-9070-ef32010a85c8"). InnerVolumeSpecName "kube-api-access-8kv6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:10:06 crc kubenswrapper[4907]: I0313 15:10:06.783983 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kv6l\" (UniqueName: \"kubernetes.io/projected/e74e8820-9ca4-4010-9070-ef32010a85c8-kube-api-access-8kv6l\") on node \"crc\" DevicePath \"\"" Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.282800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" event={"ID":"e74e8820-9ca4-4010-9070-ef32010a85c8","Type":"ContainerDied","Data":"db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d"} Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.282861 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db3872fd4c5e6ef8fbd9653013661910abc918fc75aec32d6e1e26db23ebf10d" Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.282872 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556910-5pmwx" Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.350184 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.357302 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556904-4qqhq"] Mar 13 15:10:07 crc kubenswrapper[4907]: I0313 15:10:07.794632 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="856cd90a-f3c2-4b46-b5eb-77e85004dbf6" path="/var/lib/kubelet/pods/856cd90a-f3c2-4b46-b5eb-77e85004dbf6/volumes" Mar 13 15:10:10 crc kubenswrapper[4907]: I0313 15:10:10.782925 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:10 crc kubenswrapper[4907]: E0313 15:10:10.783997 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:25 crc kubenswrapper[4907]: I0313 15:10:25.782753 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:25 crc kubenswrapper[4907]: E0313 15:10:25.783392 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:26 crc kubenswrapper[4907]: I0313 15:10:26.345522 4907 scope.go:117] "RemoveContainer" containerID="6b49bfb9250df806b9fbdd020bba3661e9f6dd127456008ebc47a00dd7745594" Mar 13 15:10:40 crc kubenswrapper[4907]: I0313 15:10:40.782777 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:40 crc kubenswrapper[4907]: E0313 15:10:40.783542 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:10:55 crc kubenswrapper[4907]: I0313 15:10:55.782783 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:10:55 crc kubenswrapper[4907]: E0313 15:10:55.785115 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:10 crc kubenswrapper[4907]: I0313 15:11:10.782957 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:10 crc kubenswrapper[4907]: E0313 15:11:10.783788 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.357840 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:20 crc kubenswrapper[4907]: E0313 15:11:20.358715 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerName="oc" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.358730 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerName="oc" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.358871 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" containerName="oc" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.359796 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.371657 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.557569 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.557643 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.557766 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.659471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.659596 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.659635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.660085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.660140 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.707130 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"redhat-marketplace-cxj9h\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:20 crc kubenswrapper[4907]: I0313 15:11:20.980245 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.537960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.786390 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:21 crc kubenswrapper[4907]: E0313 15:11:21.786923 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.940628 4907 generic.go:334] "Generic (PLEG): container finished" podID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" exitCode=0 Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.940671 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d"} Mar 13 15:11:21 crc kubenswrapper[4907]: I0313 15:11:21.940717 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerStarted","Data":"5464248f8a1cc4999ddd5ec071572b52cc1accb86479d7ff5b0ea37426d14fde"} Mar 13 15:11:22 crc kubenswrapper[4907]: I0313 15:11:22.948931 4907 generic.go:334] "Generic (PLEG): container finished" podID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" exitCode=0 Mar 13 15:11:22 crc kubenswrapper[4907]: I0313 15:11:22.948983 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c"} Mar 13 15:11:23 crc kubenswrapper[4907]: I0313 15:11:23.958861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerStarted","Data":"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364"} Mar 13 15:11:23 crc kubenswrapper[4907]: I0313 15:11:23.982721 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cxj9h" podStartSLOduration=2.495653829 podStartE2EDuration="3.982700466s" podCreationTimestamp="2026-03-13 15:11:20 +0000 UTC" firstStartedPulling="2026-03-13 15:11:21.942042095 +0000 UTC m=+3980.841829784" lastFinishedPulling="2026-03-13 15:11:23.429088722 +0000 UTC m=+3982.328876421" observedRunningTime="2026-03-13 15:11:23.978558653 +0000 UTC m=+3982.878346342" watchObservedRunningTime="2026-03-13 15:11:23.982700466 +0000 UTC m=+3982.882488155" Mar 13 15:11:30 crc kubenswrapper[4907]: I0313 15:11:30.980817 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:30 crc kubenswrapper[4907]: I0313 15:11:30.981369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:31 crc kubenswrapper[4907]: I0313 15:11:31.044453 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:31 crc kubenswrapper[4907]: I0313 15:11:31.105050 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:31 crc kubenswrapper[4907]: I0313 15:11:31.287183 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:32 crc kubenswrapper[4907]: I0313 15:11:32.782682 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:32 crc kubenswrapper[4907]: E0313 15:11:32.782944 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:11:33 crc kubenswrapper[4907]: I0313 15:11:33.020042 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cxj9h" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" containerID="cri-o://78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" gracePeriod=2 Mar 13 15:11:33 crc kubenswrapper[4907]: I0313 15:11:33.979858 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030897 4907 generic.go:334] "Generic (PLEG): container finished" podID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" exitCode=0 Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030942 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364"} Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030972 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cxj9h" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030988 4907 scope.go:117] "RemoveContainer" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.030974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cxj9h" event={"ID":"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40","Type":"ContainerDied","Data":"5464248f8a1cc4999ddd5ec071572b52cc1accb86479d7ff5b0ea37426d14fde"} Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.048937 4907 scope.go:117] "RemoveContainer" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.069504 4907 scope.go:117] "RemoveContainer" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.089958 4907 scope.go:117] "RemoveContainer" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" Mar 13 15:11:34 crc kubenswrapper[4907]: E0313 15:11:34.090461 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364\": container with ID starting with 78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364 not found: ID does not exist" containerID="78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.090490 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364"} err="failed to get container status \"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364\": rpc error: code = NotFound desc = could not find container \"78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364\": container with ID starting with 78745dd5b00ec2ec5e49e3d60ba52f4f1eb321ccddaa4efe0e88a95ab80f8364 not found: ID does not exist" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.090515 4907 scope.go:117] "RemoveContainer" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" Mar 13 15:11:34 crc kubenswrapper[4907]: E0313 15:11:34.091008 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c\": container with ID starting with 59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c not found: ID does not exist" containerID="59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.091042 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c"} err="failed to get container status \"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c\": rpc error: code = NotFound desc = could not find container \"59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c\": container with ID starting with 59fe9c3ca129487a9017d04f24aff47275d2cc5ba27e2667acd811f849f3d50c not found: ID does not exist" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.091062 4907 scope.go:117] "RemoveContainer" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" Mar 13 15:11:34 crc kubenswrapper[4907]: E0313 15:11:34.091358 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d\": container with ID starting with 7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d not found: ID does not exist" containerID="7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.091385 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d"} err="failed to get container status \"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d\": rpc error: code = NotFound desc = could not find container \"7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d\": container with ID starting with 7d419ed4bd94d03f181c4d03b47d6360d106e83f89d0f80631b8c98942f0811d not found: ID does not exist" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.145776 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") pod \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.145927 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") pod \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.145953 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") pod \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\" (UID: \"dd7cb0e4-ac20-4ff7-8b04-d40718c41d40\") " Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.147743 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities" (OuterVolumeSpecName: "utilities") pod "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" (UID: "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.151406 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b" (OuterVolumeSpecName: "kube-api-access-nzn6b") pod "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" (UID: "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40"). InnerVolumeSpecName "kube-api-access-nzn6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.178154 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" (UID: "dd7cb0e4-ac20-4ff7-8b04-d40718c41d40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.248130 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.248175 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzn6b\" (UniqueName: \"kubernetes.io/projected/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-kube-api-access-nzn6b\") on node \"crc\" DevicePath \"\"" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.248187 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.368976 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:34 crc kubenswrapper[4907]: I0313 15:11:34.374786 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cxj9h"] Mar 13 15:11:35 crc kubenswrapper[4907]: I0313 15:11:35.795525 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" path="/var/lib/kubelet/pods/dd7cb0e4-ac20-4ff7-8b04-d40718c41d40/volumes" Mar 13 15:11:46 crc kubenswrapper[4907]: I0313 15:11:46.782051 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:11:46 crc kubenswrapper[4907]: E0313 15:11:46.782827 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.164466 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:12:00 crc kubenswrapper[4907]: E0313 15:12:00.165563 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-content" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165582 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-content" Mar 13 15:12:00 crc kubenswrapper[4907]: E0313 15:12:00.165599 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165608 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" Mar 13 15:12:00 crc kubenswrapper[4907]: E0313 15:12:00.165620 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-utilities" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165629 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="extract-utilities" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.165846 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd7cb0e4-ac20-4ff7-8b04-d40718c41d40" containerName="registry-server" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.166642 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.169168 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.169203 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.170428 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.178822 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.240577 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"auto-csr-approver-29556912-smqnn\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.342473 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"auto-csr-approver-29556912-smqnn\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.368143 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"auto-csr-approver-29556912-smqnn\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.485613 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:00 crc kubenswrapper[4907]: I0313 15:12:00.941230 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:12:01 crc kubenswrapper[4907]: I0313 15:12:01.216992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556912-smqnn" event={"ID":"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9","Type":"ContainerStarted","Data":"8b02e5696714ca6118ebedfe0bc75d417d937d6760d4c0c0b82ce55f9a51fdba"} Mar 13 15:12:01 crc kubenswrapper[4907]: I0313 15:12:01.792419 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:12:02 crc kubenswrapper[4907]: I0313 15:12:02.223300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d"} Mar 13 15:12:03 crc kubenswrapper[4907]: I0313 15:12:03.230478 4907 generic.go:334] "Generic (PLEG): container finished" podID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerID="eb18542ee2a2c90ab854737b99c04f55fb12626f0e4778cc4e2869a3c3a6cd0f" exitCode=0 Mar 13 15:12:03 crc kubenswrapper[4907]: I0313 15:12:03.230529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556912-smqnn" event={"ID":"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9","Type":"ContainerDied","Data":"eb18542ee2a2c90ab854737b99c04f55fb12626f0e4778cc4e2869a3c3a6cd0f"} Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.567963 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.616561 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") pod \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\" (UID: \"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9\") " Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.623193 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs" (OuterVolumeSpecName: "kube-api-access-wg9qs") pod "449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" (UID: "449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9"). InnerVolumeSpecName "kube-api-access-wg9qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:12:04 crc kubenswrapper[4907]: I0313 15:12:04.718299 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg9qs\" (UniqueName: \"kubernetes.io/projected/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9-kube-api-access-wg9qs\") on node \"crc\" DevicePath \"\"" Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.249947 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556912-smqnn" event={"ID":"449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9","Type":"ContainerDied","Data":"8b02e5696714ca6118ebedfe0bc75d417d937d6760d4c0c0b82ce55f9a51fdba"} Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.249987 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b02e5696714ca6118ebedfe0bc75d417d937d6760d4c0c0b82ce55f9a51fdba" Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.249992 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556912-smqnn" Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.637935 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.647847 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556906-kqncj"] Mar 13 15:12:05 crc kubenswrapper[4907]: I0313 15:12:05.792499 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a501d2e-0774-414e-8afb-39a9a9d8b731" path="/var/lib/kubelet/pods/7a501d2e-0774-414e-8afb-39a9a9d8b731/volumes" Mar 13 15:12:26 crc kubenswrapper[4907]: I0313 15:12:26.477165 4907 scope.go:117] "RemoveContainer" containerID="c0f40228828ee5195fb41b365b21febab7b0feb3cf02e777e21a472c7ba50dbb" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.154281 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:14:00 crc kubenswrapper[4907]: E0313 15:14:00.156858 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerName="oc" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.156992 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerName="oc" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.157251 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" containerName="oc" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.157938 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.161545 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.162270 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.162790 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.163293 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.331227 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"auto-csr-approver-29556914-98b9m\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.433218 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"auto-csr-approver-29556914-98b9m\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.453816 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"auto-csr-approver-29556914-98b9m\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.498461 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.913090 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:14:00 crc kubenswrapper[4907]: I0313 15:14:00.922726 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:14:01 crc kubenswrapper[4907]: I0313 15:14:01.157204 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556914-98b9m" event={"ID":"561da6ce-3fe7-4e35-97f0-6a39582d3d4b","Type":"ContainerStarted","Data":"01eb835bc62db5d86ae4c0f5cf722d7dc8f945ba3bb95e0ff8b3af8bb49d05e1"} Mar 13 15:14:03 crc kubenswrapper[4907]: I0313 15:14:03.171976 4907 generic.go:334] "Generic (PLEG): container finished" podID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerID="692bff8b244c6ed0093ba50083b6fcff2e9a78403c7f3e3cc6626818a66971bb" exitCode=0 Mar 13 15:14:03 crc kubenswrapper[4907]: I0313 15:14:03.172172 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556914-98b9m" event={"ID":"561da6ce-3fe7-4e35-97f0-6a39582d3d4b","Type":"ContainerDied","Data":"692bff8b244c6ed0093ba50083b6fcff2e9a78403c7f3e3cc6626818a66971bb"} Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.459985 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.593206 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") pod \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\" (UID: \"561da6ce-3fe7-4e35-97f0-6a39582d3d4b\") " Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.598963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk" (OuterVolumeSpecName: "kube-api-access-j7vqk") pod "561da6ce-3fe7-4e35-97f0-6a39582d3d4b" (UID: "561da6ce-3fe7-4e35-97f0-6a39582d3d4b"). InnerVolumeSpecName "kube-api-access-j7vqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:14:04 crc kubenswrapper[4907]: I0313 15:14:04.694637 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7vqk\" (UniqueName: \"kubernetes.io/projected/561da6ce-3fe7-4e35-97f0-6a39582d3d4b-kube-api-access-j7vqk\") on node \"crc\" DevicePath \"\"" Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.185101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556914-98b9m" event={"ID":"561da6ce-3fe7-4e35-97f0-6a39582d3d4b","Type":"ContainerDied","Data":"01eb835bc62db5d86ae4c0f5cf722d7dc8f945ba3bb95e0ff8b3af8bb49d05e1"} Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.185497 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01eb835bc62db5d86ae4c0f5cf722d7dc8f945ba3bb95e0ff8b3af8bb49d05e1" Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.185157 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556914-98b9m" Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.526841 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.531609 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556908-lsqcs"] Mar 13 15:14:05 crc kubenswrapper[4907]: I0313 15:14:05.792557 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="555f4e30-b850-428b-a0fc-e15af18a58be" path="/var/lib/kubelet/pods/555f4e30-b850-428b-a0fc-e15af18a58be/volumes" Mar 13 15:14:18 crc kubenswrapper[4907]: I0313 15:14:18.041630 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:14:18 crc kubenswrapper[4907]: I0313 15:14:18.043261 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:14:26 crc kubenswrapper[4907]: I0313 15:14:26.576481 4907 scope.go:117] "RemoveContainer" containerID="b199bdf3c496349ef8f75e1af0202d9bd0ccc18b9b13016f199b8123d6045267" Mar 13 15:14:48 crc kubenswrapper[4907]: I0313 15:14:48.042189 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:14:48 crc kubenswrapper[4907]: I0313 15:14:48.042831 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.029957 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:14:58 crc kubenswrapper[4907]: E0313 15:14:58.035058 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerName="oc" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.035105 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerName="oc" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.035467 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" containerName="oc" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.036779 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.056796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.159790 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.159927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.160097 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261295 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261340 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.261844 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.281454 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"community-operators-jpjgk\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.369338 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:14:58 crc kubenswrapper[4907]: I0313 15:14:58.652082 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:14:59 crc kubenswrapper[4907]: I0313 15:14:59.577463 4907 generic.go:334] "Generic (PLEG): container finished" podID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" exitCode=0 Mar 13 15:14:59 crc kubenswrapper[4907]: I0313 15:14:59.577520 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497"} Mar 13 15:14:59 crc kubenswrapper[4907]: I0313 15:14:59.577551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerStarted","Data":"1e8f5904cbc6648ae855379c24f68a418662b1fb626449b7dd05f4af7ab877dc"} Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.151615 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.153066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.155363 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.155496 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.159658 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.189473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.190267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.190360 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.291041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.291108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.291171 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.292256 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.304625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.307974 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"collect-profiles-29556915-jtzgq\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.476986 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.596960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerStarted","Data":"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b"} Mar 13 15:15:00 crc kubenswrapper[4907]: I0313 15:15:00.898129 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 15:15:00 crc kubenswrapper[4907]: W0313 15:15:00.908571 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fb6a3c2_e8df_47a4_9ae7_2ee38842051a.slice/crio-8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83 WatchSource:0}: Error finding container 8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83: Status 404 returned error can't find the container with id 8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83 Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.608666 4907 generic.go:334] "Generic (PLEG): container finished" podID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" exitCode=0 Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.608776 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b"} Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.612313 4907 generic.go:334] "Generic (PLEG): container finished" podID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerID="f8d1778bda06ccfd24371ed557383017005b23e6277f212ccf01bbaf98e64b94" exitCode=0 Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.612368 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" event={"ID":"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a","Type":"ContainerDied","Data":"f8d1778bda06ccfd24371ed557383017005b23e6277f212ccf01bbaf98e64b94"} Mar 13 15:15:01 crc kubenswrapper[4907]: I0313 15:15:01.612403 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" event={"ID":"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a","Type":"ContainerStarted","Data":"8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83"} Mar 13 15:15:02 crc kubenswrapper[4907]: I0313 15:15:02.621257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerStarted","Data":"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30"} Mar 13 15:15:02 crc kubenswrapper[4907]: I0313 15:15:02.643180 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jpjgk" podStartSLOduration=2.119638662 podStartE2EDuration="4.643163727s" podCreationTimestamp="2026-03-13 15:14:58 +0000 UTC" firstStartedPulling="2026-03-13 15:14:59.580836934 +0000 UTC m=+4198.480624663" lastFinishedPulling="2026-03-13 15:15:02.104362039 +0000 UTC m=+4201.004149728" observedRunningTime="2026-03-13 15:15:02.638380766 +0000 UTC m=+4201.538168475" watchObservedRunningTime="2026-03-13 15:15:02.643163727 +0000 UTC m=+4201.542951416" Mar 13 15:15:02 crc kubenswrapper[4907]: I0313 15:15:02.882035 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.041270 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") pod \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.041379 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") pod \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.041399 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") pod \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\" (UID: \"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a\") " Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.042468 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume" (OuterVolumeSpecName: "config-volume") pod "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" (UID: "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.046715 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" (UID: "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.053702 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86" (OuterVolumeSpecName: "kube-api-access-fxh86") pod "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" (UID: "1fb6a3c2-e8df-47a4-9ae7-2ee38842051a"). InnerVolumeSpecName "kube-api-access-fxh86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.143196 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.143255 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxh86\" (UniqueName: \"kubernetes.io/projected/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-kube-api-access-fxh86\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.143273 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.628056 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.628077 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq" event={"ID":"1fb6a3c2-e8df-47a4-9ae7-2ee38842051a","Type":"ContainerDied","Data":"8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83"} Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.628130 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8763e76268e6fdebf8c7827617d0774f78f5d7e8ce50c07417aa72eaaa7a2d83" Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.967837 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 15:15:03 crc kubenswrapper[4907]: I0313 15:15:03.972849 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556870-rdhxg"] Mar 13 15:15:05 crc kubenswrapper[4907]: I0313 15:15:05.790693 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af9d29dd-6749-4fcd-82a8-ea528d832fcc" path="/var/lib/kubelet/pods/af9d29dd-6749-4fcd-82a8-ea528d832fcc/volumes" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.370171 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.370539 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.432807 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.702004 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:08 crc kubenswrapper[4907]: I0313 15:15:08.747163 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:15:10 crc kubenswrapper[4907]: I0313 15:15:10.681513 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jpjgk" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" containerID="cri-o://2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" gracePeriod=2 Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.113669 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.257397 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") pod \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.257602 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") pod \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.257710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") pod \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\" (UID: \"db74eac8-6eb9-4abc-8bd4-05a7c4de409d\") " Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.258996 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities" (OuterVolumeSpecName: "utilities") pod "db74eac8-6eb9-4abc-8bd4-05a7c4de409d" (UID: "db74eac8-6eb9-4abc-8bd4-05a7c4de409d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.359532 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.412309 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5" (OuterVolumeSpecName: "kube-api-access-bkfd5") pod "db74eac8-6eb9-4abc-8bd4-05a7c4de409d" (UID: "db74eac8-6eb9-4abc-8bd4-05a7c4de409d"). InnerVolumeSpecName "kube-api-access-bkfd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.460704 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkfd5\" (UniqueName: \"kubernetes.io/projected/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-kube-api-access-bkfd5\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692843 4907 generic.go:334] "Generic (PLEG): container finished" podID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" exitCode=0 Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692906 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30"} Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jpjgk" event={"ID":"db74eac8-6eb9-4abc-8bd4-05a7c4de409d","Type":"ContainerDied","Data":"1e8f5904cbc6648ae855379c24f68a418662b1fb626449b7dd05f4af7ab877dc"} Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.692983 4907 scope.go:117] "RemoveContainer" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.693128 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jpjgk" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.721917 4907 scope.go:117] "RemoveContainer" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.742676 4907 scope.go:117] "RemoveContainer" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.778174 4907 scope.go:117] "RemoveContainer" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" Mar 13 15:15:11 crc kubenswrapper[4907]: E0313 15:15:11.778723 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30\": container with ID starting with 2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30 not found: ID does not exist" containerID="2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.778762 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30"} err="failed to get container status \"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30\": rpc error: code = NotFound desc = could not find container \"2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30\": container with ID starting with 2832afb1ca9f1fcdd05a62540f3537314948bd51bef218b6399c500e2dfcca30 not found: ID does not exist" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.778789 4907 scope.go:117] "RemoveContainer" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" Mar 13 15:15:11 crc kubenswrapper[4907]: E0313 15:15:11.779224 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b\": container with ID starting with 43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b not found: ID does not exist" containerID="43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.779277 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b"} err="failed to get container status \"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b\": rpc error: code = NotFound desc = could not find container \"43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b\": container with ID starting with 43c417bf05f6d13ba2647567c1378035b8d2b1457c7230a97bd627f1015c7e9b not found: ID does not exist" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.779310 4907 scope.go:117] "RemoveContainer" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" Mar 13 15:15:11 crc kubenswrapper[4907]: E0313 15:15:11.779669 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497\": container with ID starting with 40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497 not found: ID does not exist" containerID="40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.779709 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497"} err="failed to get container status \"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497\": rpc error: code = NotFound desc = could not find container \"40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497\": container with ID starting with 40cf43ff58a1d7d18d1fe371ca4cdf1f9b55f77c9ec021d33f6565bbdaadc497 not found: ID does not exist" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.801640 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db74eac8-6eb9-4abc-8bd4-05a7c4de409d" (UID: "db74eac8-6eb9-4abc-8bd4-05a7c4de409d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:15:11 crc kubenswrapper[4907]: I0313 15:15:11.866704 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db74eac8-6eb9-4abc-8bd4-05a7c4de409d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:15:12 crc kubenswrapper[4907]: I0313 15:15:12.030580 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:15:12 crc kubenswrapper[4907]: I0313 15:15:12.037742 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jpjgk"] Mar 13 15:15:13 crc kubenswrapper[4907]: I0313 15:15:13.795183 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" path="/var/lib/kubelet/pods/db74eac8-6eb9-4abc-8bd4-05a7c4de409d/volumes" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.041731 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.042123 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.042184 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.043153 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.043278 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d" gracePeriod=600 Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.740675 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d" exitCode=0 Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.740769 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d"} Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.741437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c"} Mar 13 15:15:18 crc kubenswrapper[4907]: I0313 15:15:18.741485 4907 scope.go:117] "RemoveContainer" containerID="caeb29efd9b0ec75cd14d4d1c0fbb54dce8073ba7639ceb605cc152e79561f8a" Mar 13 15:15:26 crc kubenswrapper[4907]: I0313 15:15:26.633480 4907 scope.go:117] "RemoveContainer" containerID="46dfd46e5c29fa7764c024392038bd85a0e04092c9a231bbd08b67969852ba5c" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.145626 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.146767 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-content" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.146786 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-content" Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.146814 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.146827 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.146856 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-utilities" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.146870 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="extract-utilities" Mar 13 15:16:00 crc kubenswrapper[4907]: E0313 15:16:00.147010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerName="collect-profiles" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147024 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerName="collect-profiles" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147253 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="db74eac8-6eb9-4abc-8bd4-05a7c4de409d" containerName="registry-server" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147281 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" containerName="collect-profiles" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.147987 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.153343 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.154008 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.154211 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.164417 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.281779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"auto-csr-approver-29556916-gpp98\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.383771 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"auto-csr-approver-29556916-gpp98\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.417159 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"auto-csr-approver-29556916-gpp98\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.478120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:00 crc kubenswrapper[4907]: I0313 15:16:00.907920 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:16:01 crc kubenswrapper[4907]: I0313 15:16:01.046708 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556916-gpp98" event={"ID":"c5c630da-de73-4f29-89f0-43c5497dd89a","Type":"ContainerStarted","Data":"93ed1baee1884167038c99f469dc1ebb5a70d1aba4262c7b4e9fae45a3e3a33c"} Mar 13 15:16:03 crc kubenswrapper[4907]: I0313 15:16:03.062105 4907 generic.go:334] "Generic (PLEG): container finished" podID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerID="348c87a146da948df0e4139184e97c7692cc71aad3ec14fe9f015dc10660b3b8" exitCode=0 Mar 13 15:16:03 crc kubenswrapper[4907]: I0313 15:16:03.062209 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556916-gpp98" event={"ID":"c5c630da-de73-4f29-89f0-43c5497dd89a","Type":"ContainerDied","Data":"348c87a146da948df0e4139184e97c7692cc71aad3ec14fe9f015dc10660b3b8"} Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.358190 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.548403 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") pod \"c5c630da-de73-4f29-89f0-43c5497dd89a\" (UID: \"c5c630da-de73-4f29-89f0-43c5497dd89a\") " Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.555706 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz" (OuterVolumeSpecName: "kube-api-access-sxmtz") pod "c5c630da-de73-4f29-89f0-43c5497dd89a" (UID: "c5c630da-de73-4f29-89f0-43c5497dd89a"). InnerVolumeSpecName "kube-api-access-sxmtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:16:04 crc kubenswrapper[4907]: I0313 15:16:04.649761 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxmtz\" (UniqueName: \"kubernetes.io/projected/c5c630da-de73-4f29-89f0-43c5497dd89a-kube-api-access-sxmtz\") on node \"crc\" DevicePath \"\"" Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.078625 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556916-gpp98" event={"ID":"c5c630da-de73-4f29-89f0-43c5497dd89a","Type":"ContainerDied","Data":"93ed1baee1884167038c99f469dc1ebb5a70d1aba4262c7b4e9fae45a3e3a33c"} Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.078674 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93ed1baee1884167038c99f469dc1ebb5a70d1aba4262c7b4e9fae45a3e3a33c" Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.078733 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556916-gpp98" Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.425082 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.433631 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556910-5pmwx"] Mar 13 15:16:05 crc kubenswrapper[4907]: I0313 15:16:05.791223 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e74e8820-9ca4-4010-9070-ef32010a85c8" path="/var/lib/kubelet/pods/e74e8820-9ca4-4010-9070-ef32010a85c8/volumes" Mar 13 15:16:26 crc kubenswrapper[4907]: I0313 15:16:26.704012 4907 scope.go:117] "RemoveContainer" containerID="6156d45b4121389ab4793e9ba61feb497d3ccb0338571d8b44490cfdec3820ce" Mar 13 15:17:18 crc kubenswrapper[4907]: I0313 15:17:18.042201 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:17:18 crc kubenswrapper[4907]: I0313 15:17:18.042988 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.675569 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:32 crc kubenswrapper[4907]: E0313 15:17:32.676593 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerName="oc" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.676611 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerName="oc" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.676841 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" containerName="oc" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.678163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.683471 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.859556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.859613 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.859638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.960776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.960836 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.960971 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.961324 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.961410 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:32 crc kubenswrapper[4907]: I0313 15:17:32.992446 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"certified-operators-k675m\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.002324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.526603 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.741560 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" exitCode=0 Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.741656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6"} Mar 13 15:17:33 crc kubenswrapper[4907]: I0313 15:17:33.741832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerStarted","Data":"4e3fd1a5082ad9dee2e764ae30a95c9a0adcb1f77f2fd1c6a9389907f1dd1779"} Mar 13 15:17:34 crc kubenswrapper[4907]: I0313 15:17:34.749255 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" exitCode=0 Mar 13 15:17:34 crc kubenswrapper[4907]: I0313 15:17:34.749422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188"} Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.757284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerStarted","Data":"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58"} Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.776028 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k675m" podStartSLOduration=2.372191954 podStartE2EDuration="3.776011351s" podCreationTimestamp="2026-03-13 15:17:32 +0000 UTC" firstStartedPulling="2026-03-13 15:17:33.74337868 +0000 UTC m=+4352.643166369" lastFinishedPulling="2026-03-13 15:17:35.147198087 +0000 UTC m=+4354.046985766" observedRunningTime="2026-03-13 15:17:35.774464169 +0000 UTC m=+4354.674251878" watchObservedRunningTime="2026-03-13 15:17:35.776011351 +0000 UTC m=+4354.675799040" Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.867149 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.868910 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:35 crc kubenswrapper[4907]: I0313 15:17:35.877033 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.000286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.000344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.000377 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.102934 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103217 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103338 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103804 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.103910 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.121366 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"redhat-operators-xhrm7\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.191336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.647592 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:36 crc kubenswrapper[4907]: I0313 15:17:36.764382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerStarted","Data":"e710e5af60dad227fb28f3f3f8736be015c34c3cbf424e8cb7767ac5bea1fdd9"} Mar 13 15:17:37 crc kubenswrapper[4907]: I0313 15:17:37.774118 4907 generic.go:334] "Generic (PLEG): container finished" podID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" exitCode=0 Mar 13 15:17:37 crc kubenswrapper[4907]: I0313 15:17:37.774166 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d"} Mar 13 15:17:39 crc kubenswrapper[4907]: I0313 15:17:39.792068 4907 generic.go:334] "Generic (PLEG): container finished" podID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" exitCode=0 Mar 13 15:17:39 crc kubenswrapper[4907]: I0313 15:17:39.792131 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958"} Mar 13 15:17:40 crc kubenswrapper[4907]: I0313 15:17:40.800497 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerStarted","Data":"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c"} Mar 13 15:17:40 crc kubenswrapper[4907]: I0313 15:17:40.818841 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xhrm7" podStartSLOduration=3.399974499 podStartE2EDuration="5.818818746s" podCreationTimestamp="2026-03-13 15:17:35 +0000 UTC" firstStartedPulling="2026-03-13 15:17:37.775720649 +0000 UTC m=+4356.675508338" lastFinishedPulling="2026-03-13 15:17:40.194564896 +0000 UTC m=+4359.094352585" observedRunningTime="2026-03-13 15:17:40.815851685 +0000 UTC m=+4359.715639374" watchObservedRunningTime="2026-03-13 15:17:40.818818746 +0000 UTC m=+4359.718606435" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.003703 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.004235 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.070688 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:43 crc kubenswrapper[4907]: I0313 15:17:43.864296 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:46 crc kubenswrapper[4907]: I0313 15:17:46.390324 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:46 crc kubenswrapper[4907]: I0313 15:17:46.391115 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:47 crc kubenswrapper[4907]: I0313 15:17:47.432516 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xhrm7" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" probeResult="failure" output=< Mar 13 15:17:47 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:17:47 crc kubenswrapper[4907]: > Mar 13 15:17:47 crc kubenswrapper[4907]: I0313 15:17:47.859561 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:47 crc kubenswrapper[4907]: I0313 15:17:47.859857 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k675m" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" containerID="cri-o://de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" gracePeriod=2 Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.042282 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.042739 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.226330 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.366602 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") pod \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.366702 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") pod \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.366728 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") pod \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\" (UID: \"7a6ef1dc-77f0-4a25-9465-f9b881023f83\") " Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.367731 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities" (OuterVolumeSpecName: "utilities") pod "7a6ef1dc-77f0-4a25-9465-f9b881023f83" (UID: "7a6ef1dc-77f0-4a25-9465-f9b881023f83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.368073 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.373988 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl" (OuterVolumeSpecName: "kube-api-access-ng4nl") pod "7a6ef1dc-77f0-4a25-9465-f9b881023f83" (UID: "7a6ef1dc-77f0-4a25-9465-f9b881023f83"). InnerVolumeSpecName "kube-api-access-ng4nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.428510 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a6ef1dc-77f0-4a25-9465-f9b881023f83" (UID: "7a6ef1dc-77f0-4a25-9465-f9b881023f83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.469225 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6ef1dc-77f0-4a25-9465-f9b881023f83-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.469252 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng4nl\" (UniqueName: \"kubernetes.io/projected/7a6ef1dc-77f0-4a25-9465-f9b881023f83-kube-api-access-ng4nl\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.858700 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" exitCode=0 Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.858780 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k675m" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.858789 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58"} Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.859459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k675m" event={"ID":"7a6ef1dc-77f0-4a25-9465-f9b881023f83","Type":"ContainerDied","Data":"4e3fd1a5082ad9dee2e764ae30a95c9a0adcb1f77f2fd1c6a9389907f1dd1779"} Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.859517 4907 scope.go:117] "RemoveContainer" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.899762 4907 scope.go:117] "RemoveContainer" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.903143 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.924257 4907 scope.go:117] "RemoveContainer" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.928928 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k675m"] Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.948807 4907 scope.go:117] "RemoveContainer" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" Mar 13 15:17:48 crc kubenswrapper[4907]: E0313 15:17:48.949324 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58\": container with ID starting with de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58 not found: ID does not exist" containerID="de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.949395 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58"} err="failed to get container status \"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58\": rpc error: code = NotFound desc = could not find container \"de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58\": container with ID starting with de1a6aa0e4c46c2959d6be7168e2b145807f3a5c7a00ecfda7923f20d6b96d58 not found: ID does not exist" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.949434 4907 scope.go:117] "RemoveContainer" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" Mar 13 15:17:48 crc kubenswrapper[4907]: E0313 15:17:48.949943 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188\": container with ID starting with 2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188 not found: ID does not exist" containerID="2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.950012 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188"} err="failed to get container status \"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188\": rpc error: code = NotFound desc = could not find container \"2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188\": container with ID starting with 2cbc3b0e8cb89dede628cdba2545a3b54133f559dd5f2d0fcaf7f50b62b92188 not found: ID does not exist" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.950047 4907 scope.go:117] "RemoveContainer" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" Mar 13 15:17:48 crc kubenswrapper[4907]: E0313 15:17:48.950359 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6\": container with ID starting with f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6 not found: ID does not exist" containerID="f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6" Mar 13 15:17:48 crc kubenswrapper[4907]: I0313 15:17:48.950394 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6"} err="failed to get container status \"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6\": rpc error: code = NotFound desc = could not find container \"f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6\": container with ID starting with f64382383e10d8bc77380b8411efc74df96b9cacb9ce301d06c4749e0ccc97b6 not found: ID does not exist" Mar 13 15:17:49 crc kubenswrapper[4907]: E0313 15:17:49.070000 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6ef1dc_77f0_4a25_9465_f9b881023f83.slice/crio-4e3fd1a5082ad9dee2e764ae30a95c9a0adcb1f77f2fd1c6a9389907f1dd1779\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6ef1dc_77f0_4a25_9465_f9b881023f83.slice\": RecentStats: unable to find data in memory cache]" Mar 13 15:17:49 crc kubenswrapper[4907]: I0313 15:17:49.791764 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" path="/var/lib/kubelet/pods/7a6ef1dc-77f0-4a25-9465-f9b881023f83/volumes" Mar 13 15:17:56 crc kubenswrapper[4907]: I0313 15:17:56.234351 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:56 crc kubenswrapper[4907]: I0313 15:17:56.276760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:56 crc kubenswrapper[4907]: I0313 15:17:56.467630 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:57 crc kubenswrapper[4907]: I0313 15:17:57.927638 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xhrm7" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" containerID="cri-o://91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" gracePeriod=2 Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.321717 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.412955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") pod \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.413022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") pod \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.413099 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") pod \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\" (UID: \"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f\") " Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.414387 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities" (OuterVolumeSpecName: "utilities") pod "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" (UID: "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.424189 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx" (OuterVolumeSpecName: "kube-api-access-m4gbx") pod "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" (UID: "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f"). InnerVolumeSpecName "kube-api-access-m4gbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.515143 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4gbx\" (UniqueName: \"kubernetes.io/projected/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-kube-api-access-m4gbx\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.515194 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.563484 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" (UID: "78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.616728 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937451 4907 generic.go:334] "Generic (PLEG): container finished" podID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" exitCode=0 Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937533 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c"} Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937574 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xhrm7" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937616 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xhrm7" event={"ID":"78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f","Type":"ContainerDied","Data":"e710e5af60dad227fb28f3f3f8736be015c34c3cbf424e8cb7767ac5bea1fdd9"} Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.937647 4907 scope.go:117] "RemoveContainer" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.973660 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.973811 4907 scope.go:117] "RemoveContainer" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" Mar 13 15:17:58 crc kubenswrapper[4907]: I0313 15:17:58.988225 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xhrm7"] Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.011280 4907 scope.go:117] "RemoveContainer" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.040273 4907 scope.go:117] "RemoveContainer" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" Mar 13 15:17:59 crc kubenswrapper[4907]: E0313 15:17:59.041504 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c\": container with ID starting with 91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c not found: ID does not exist" containerID="91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.041546 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c"} err="failed to get container status \"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c\": rpc error: code = NotFound desc = could not find container \"91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c\": container with ID starting with 91ebf771b78bb7a2fe3a8359fd7a3263f5fdd1c2032beba161125b81cdf81d3c not found: ID does not exist" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.041573 4907 scope.go:117] "RemoveContainer" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" Mar 13 15:17:59 crc kubenswrapper[4907]: E0313 15:17:59.042070 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958\": container with ID starting with 21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958 not found: ID does not exist" containerID="21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.042112 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958"} err="failed to get container status \"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958\": rpc error: code = NotFound desc = could not find container \"21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958\": container with ID starting with 21f6ddb10036b57ffc641c1c43e81015c667a267b0b5f356fdd1eeded7b6a958 not found: ID does not exist" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.042140 4907 scope.go:117] "RemoveContainer" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" Mar 13 15:17:59 crc kubenswrapper[4907]: E0313 15:17:59.042520 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d\": container with ID starting with cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d not found: ID does not exist" containerID="cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.042573 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d"} err="failed to get container status \"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d\": rpc error: code = NotFound desc = could not find container \"cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d\": container with ID starting with cdd1adc8cc682b604b6e997800c49a6fbb65e62d6431a882b09277e919d44d4d not found: ID does not exist" Mar 13 15:17:59 crc kubenswrapper[4907]: I0313 15:17:59.794493 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" path="/var/lib/kubelet/pods/78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f/volumes" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.154960 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155832 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155855 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155910 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155926 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155954 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155967 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.155985 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.155998 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="extract-utilities" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.156018 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156030 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: E0313 15:18:00.156057 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156069 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="extract-content" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="78255f5a-d6ed-45c3-ade0-4ff6f01c7a0f" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.156346 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6ef1dc-77f0-4a25-9465-f9b881023f83" containerName="registry-server" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.157167 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.160631 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.160832 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.160993 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.192942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.237709 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"auto-csr-approver-29556918-cs4dh\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.338844 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"auto-csr-approver-29556918-cs4dh\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.356234 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"auto-csr-approver-29556918-cs4dh\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.501374 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.902714 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:18:00 crc kubenswrapper[4907]: I0313 15:18:00.960608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" event={"ID":"fdd030a7-3ebb-4fbf-91f2-2ab96e526252","Type":"ContainerStarted","Data":"51efcbdac2654ebcc5d956fb9add4884a6d192c679448f44a80b43591fcf224b"} Mar 13 15:18:02 crc kubenswrapper[4907]: I0313 15:18:02.985500 4907 generic.go:334] "Generic (PLEG): container finished" podID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerID="e984813306df1ab82de9a9dd04affc9e989fc2092eb575a81e8fc5794b033f1a" exitCode=0 Mar 13 15:18:02 crc kubenswrapper[4907]: I0313 15:18:02.985664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" event={"ID":"fdd030a7-3ebb-4fbf-91f2-2ab96e526252","Type":"ContainerDied","Data":"e984813306df1ab82de9a9dd04affc9e989fc2092eb575a81e8fc5794b033f1a"} Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.285676 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.395834 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") pod \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\" (UID: \"fdd030a7-3ebb-4fbf-91f2-2ab96e526252\") " Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.400849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb" (OuterVolumeSpecName: "kube-api-access-6m4sb") pod "fdd030a7-3ebb-4fbf-91f2-2ab96e526252" (UID: "fdd030a7-3ebb-4fbf-91f2-2ab96e526252"). InnerVolumeSpecName "kube-api-access-6m4sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:18:04 crc kubenswrapper[4907]: I0313 15:18:04.497475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m4sb\" (UniqueName: \"kubernetes.io/projected/fdd030a7-3ebb-4fbf-91f2-2ab96e526252-kube-api-access-6m4sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.002147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" event={"ID":"fdd030a7-3ebb-4fbf-91f2-2ab96e526252","Type":"ContainerDied","Data":"51efcbdac2654ebcc5d956fb9add4884a6d192c679448f44a80b43591fcf224b"} Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.002202 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51efcbdac2654ebcc5d956fb9add4884a6d192c679448f44a80b43591fcf224b" Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.002266 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556918-cs4dh" Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.368274 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.376970 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556912-smqnn"] Mar 13 15:18:05 crc kubenswrapper[4907]: I0313 15:18:05.790337 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9" path="/var/lib/kubelet/pods/449e0ca3-4756-4ec2-8ea2-b2b6fbe52ca9/volumes" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.041665 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.042293 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.042355 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.043144 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:18:18 crc kubenswrapper[4907]: I0313 15:18:18.043245 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" gracePeriod=600 Mar 13 15:18:18 crc kubenswrapper[4907]: E0313 15:18:18.178096 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.120998 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" exitCode=0 Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.121046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c"} Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.121088 4907 scope.go:117] "RemoveContainer" containerID="36ef7b96f3457426d961c14798528949d4c51116631e09ea3be0c17a7c07b91d" Mar 13 15:18:19 crc kubenswrapper[4907]: I0313 15:18:19.121841 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:19 crc kubenswrapper[4907]: E0313 15:18:19.122108 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:26 crc kubenswrapper[4907]: I0313 15:18:26.807193 4907 scope.go:117] "RemoveContainer" containerID="eb18542ee2a2c90ab854737b99c04f55fb12626f0e4778cc4e2869a3c3a6cd0f" Mar 13 15:18:30 crc kubenswrapper[4907]: I0313 15:18:30.782751 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:30 crc kubenswrapper[4907]: E0313 15:18:30.783980 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:45 crc kubenswrapper[4907]: I0313 15:18:45.782847 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:45 crc kubenswrapper[4907]: E0313 15:18:45.783718 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:18:58 crc kubenswrapper[4907]: I0313 15:18:58.783088 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:18:58 crc kubenswrapper[4907]: E0313 15:18:58.784170 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:13 crc kubenswrapper[4907]: I0313 15:19:13.782153 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:13 crc kubenswrapper[4907]: E0313 15:19:13.782938 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:27 crc kubenswrapper[4907]: I0313 15:19:27.783013 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:27 crc kubenswrapper[4907]: E0313 15:19:27.783863 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:39 crc kubenswrapper[4907]: I0313 15:19:39.782958 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:39 crc kubenswrapper[4907]: E0313 15:19:39.783827 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:19:52 crc kubenswrapper[4907]: I0313 15:19:52.782783 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:19:52 crc kubenswrapper[4907]: E0313 15:19:52.783497 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.146158 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:20:00 crc kubenswrapper[4907]: E0313 15:20:00.147080 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerName="oc" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.147096 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerName="oc" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.147262 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" containerName="oc" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.147826 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.150925 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.151489 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.151582 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.151924 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.260669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"auto-csr-approver-29556920-lrh8k\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.362730 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"auto-csr-approver-29556920-lrh8k\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.381426 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"auto-csr-approver-29556920-lrh8k\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.469447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.681123 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:20:00 crc kubenswrapper[4907]: I0313 15:20:00.695127 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:20:01 crc kubenswrapper[4907]: I0313 15:20:01.241714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" event={"ID":"925bf3a9-9978-49a9-8ac3-7c92a928adb7","Type":"ContainerStarted","Data":"9f0ae93c37a3701544ad75512f3f6ea72f66a0bf17bc5937952fb61e324abb61"} Mar 13 15:20:02 crc kubenswrapper[4907]: I0313 15:20:02.251218 4907 generic.go:334] "Generic (PLEG): container finished" podID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerID="e5083f5e7cb4434ec8bff6d3c9df8bdd433195c0fc9dbe495454864580276a9b" exitCode=0 Mar 13 15:20:02 crc kubenswrapper[4907]: I0313 15:20:02.251308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" event={"ID":"925bf3a9-9978-49a9-8ac3-7c92a928adb7","Type":"ContainerDied","Data":"e5083f5e7cb4434ec8bff6d3c9df8bdd433195c0fc9dbe495454864580276a9b"} Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.563074 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.612918 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") pod \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\" (UID: \"925bf3a9-9978-49a9-8ac3-7c92a928adb7\") " Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.705122 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd" (OuterVolumeSpecName: "kube-api-access-ltdsd") pod "925bf3a9-9978-49a9-8ac3-7c92a928adb7" (UID: "925bf3a9-9978-49a9-8ac3-7c92a928adb7"). InnerVolumeSpecName "kube-api-access-ltdsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:20:03 crc kubenswrapper[4907]: I0313 15:20:03.714661 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltdsd\" (UniqueName: \"kubernetes.io/projected/925bf3a9-9978-49a9-8ac3-7c92a928adb7-kube-api-access-ltdsd\") on node \"crc\" DevicePath \"\"" Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.265199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" event={"ID":"925bf3a9-9978-49a9-8ac3-7c92a928adb7","Type":"ContainerDied","Data":"9f0ae93c37a3701544ad75512f3f6ea72f66a0bf17bc5937952fb61e324abb61"} Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.265479 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f0ae93c37a3701544ad75512f3f6ea72f66a0bf17bc5937952fb61e324abb61" Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.265292 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556920-lrh8k" Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.624039 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:20:04 crc kubenswrapper[4907]: I0313 15:20:04.629318 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556914-98b9m"] Mar 13 15:20:05 crc kubenswrapper[4907]: I0313 15:20:05.782160 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:05 crc kubenswrapper[4907]: E0313 15:20:05.782378 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:05 crc kubenswrapper[4907]: I0313 15:20:05.795424 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561da6ce-3fe7-4e35-97f0-6a39582d3d4b" path="/var/lib/kubelet/pods/561da6ce-3fe7-4e35-97f0-6a39582d3d4b/volumes" Mar 13 15:20:18 crc kubenswrapper[4907]: I0313 15:20:18.782996 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:18 crc kubenswrapper[4907]: E0313 15:20:18.784350 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:27 crc kubenswrapper[4907]: I0313 15:20:27.003448 4907 scope.go:117] "RemoveContainer" containerID="692bff8b244c6ed0093ba50083b6fcff2e9a78403c7f3e3cc6626818a66971bb" Mar 13 15:20:32 crc kubenswrapper[4907]: I0313 15:20:32.782391 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:32 crc kubenswrapper[4907]: E0313 15:20:32.783127 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:43 crc kubenswrapper[4907]: I0313 15:20:43.782324 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:43 crc kubenswrapper[4907]: E0313 15:20:43.784271 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:20:57 crc kubenswrapper[4907]: I0313 15:20:57.782508 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:20:57 crc kubenswrapper[4907]: E0313 15:20:57.783246 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:11 crc kubenswrapper[4907]: I0313 15:21:11.786666 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:11 crc kubenswrapper[4907]: E0313 15:21:11.787455 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:22 crc kubenswrapper[4907]: I0313 15:21:22.781721 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:22 crc kubenswrapper[4907]: E0313 15:21:22.782562 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:36 crc kubenswrapper[4907]: I0313 15:21:36.783553 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:36 crc kubenswrapper[4907]: E0313 15:21:36.784507 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.743213 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:21:48 crc kubenswrapper[4907]: E0313 15:21:48.744139 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerName="oc" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.744155 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerName="oc" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.744324 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" containerName="oc" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.745485 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.762253 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.781863 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:21:48 crc kubenswrapper[4907]: E0313 15:21:48.782232 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.871100 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.871277 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.871343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972360 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972435 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972921 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.972970 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:48 crc kubenswrapper[4907]: I0313 15:21:48.994940 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"redhat-marketplace-qtts8\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:49 crc kubenswrapper[4907]: I0313 15:21:49.064182 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:49 crc kubenswrapper[4907]: I0313 15:21:49.562767 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:21:50 crc kubenswrapper[4907]: I0313 15:21:50.040227 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" exitCode=0 Mar 13 15:21:50 crc kubenswrapper[4907]: I0313 15:21:50.040284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948"} Mar 13 15:21:50 crc kubenswrapper[4907]: I0313 15:21:50.040320 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerStarted","Data":"6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12"} Mar 13 15:21:51 crc kubenswrapper[4907]: I0313 15:21:51.050141 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" exitCode=0 Mar 13 15:21:51 crc kubenswrapper[4907]: I0313 15:21:51.050297 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a"} Mar 13 15:21:52 crc kubenswrapper[4907]: I0313 15:21:52.058388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerStarted","Data":"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04"} Mar 13 15:21:52 crc kubenswrapper[4907]: I0313 15:21:52.081105 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qtts8" podStartSLOduration=2.659200975 podStartE2EDuration="4.081085389s" podCreationTimestamp="2026-03-13 15:21:48 +0000 UTC" firstStartedPulling="2026-03-13 15:21:50.042803523 +0000 UTC m=+4608.942591212" lastFinishedPulling="2026-03-13 15:21:51.464687937 +0000 UTC m=+4610.364475626" observedRunningTime="2026-03-13 15:21:52.075749643 +0000 UTC m=+4610.975537342" watchObservedRunningTime="2026-03-13 15:21:52.081085389 +0000 UTC m=+4610.980873078" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.064801 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.065302 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.108649 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.152116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:21:59 crc kubenswrapper[4907]: I0313 15:21:59.343276 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.136830 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.138130 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.143512 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.143555 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.143571 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.145162 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.161264 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"auto-csr-approver-29556922-b7qtk\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.263088 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"auto-csr-approver-29556922-b7qtk\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.283936 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"auto-csr-approver-29556922-b7qtk\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.458839 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.783180 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:00 crc kubenswrapper[4907]: E0313 15:22:00.783777 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:00 crc kubenswrapper[4907]: I0313 15:22:00.879917 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.121222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerStarted","Data":"0e6b3e4cee3836f04db13b82a09d26c6bd36e7d4ef4e9ebe75b7a51fdd3a4b97"} Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.121388 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qtts8" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" containerID="cri-o://011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" gracePeriod=2 Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.659362 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.684827 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") pod \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.684885 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") pod \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.685022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") pod \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\" (UID: \"8ce4bcd2-1af1-43b6-8e59-d470e9089686\") " Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.685982 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities" (OuterVolumeSpecName: "utilities") pod "8ce4bcd2-1af1-43b6-8e59-d470e9089686" (UID: "8ce4bcd2-1af1-43b6-8e59-d470e9089686"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.691592 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq" (OuterVolumeSpecName: "kube-api-access-56pfq") pod "8ce4bcd2-1af1-43b6-8e59-d470e9089686" (UID: "8ce4bcd2-1af1-43b6-8e59-d470e9089686"). InnerVolumeSpecName "kube-api-access-56pfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.714617 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ce4bcd2-1af1-43b6-8e59-d470e9089686" (UID: "8ce4bcd2-1af1-43b6-8e59-d470e9089686"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.786118 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.786160 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ce4bcd2-1af1-43b6-8e59-d470e9089686-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:01 crc kubenswrapper[4907]: I0313 15:22:01.786178 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56pfq\" (UniqueName: \"kubernetes.io/projected/8ce4bcd2-1af1-43b6-8e59-d470e9089686-kube-api-access-56pfq\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135141 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" exitCode=0 Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04"} Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135703 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qtts8" event={"ID":"8ce4bcd2-1af1-43b6-8e59-d470e9089686","Type":"ContainerDied","Data":"6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12"} Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135728 4907 scope.go:117] "RemoveContainer" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.135245 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qtts8" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.141946 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerStarted","Data":"2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519"} Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.156784 4907 scope.go:117] "RemoveContainer" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.158680 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.164769 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qtts8"] Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.174177 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" podStartSLOduration=1.223280165 podStartE2EDuration="2.174157671s" podCreationTimestamp="2026-03-13 15:22:00 +0000 UTC" firstStartedPulling="2026-03-13 15:22:00.903750218 +0000 UTC m=+4619.803537907" lastFinishedPulling="2026-03-13 15:22:01.854627724 +0000 UTC m=+4620.754415413" observedRunningTime="2026-03-13 15:22:02.169751791 +0000 UTC m=+4621.069539480" watchObservedRunningTime="2026-03-13 15:22:02.174157671 +0000 UTC m=+4621.073945360" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.182978 4907 scope.go:117] "RemoveContainer" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.206719 4907 scope.go:117] "RemoveContainer" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" Mar 13 15:22:02 crc kubenswrapper[4907]: E0313 15:22:02.207090 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04\": container with ID starting with 011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04 not found: ID does not exist" containerID="011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207117 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04"} err="failed to get container status \"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04\": rpc error: code = NotFound desc = could not find container \"011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04\": container with ID starting with 011e7e094e32e738642f126d721a01c4dd2932844c6b73664a0cb1d95cb88f04 not found: ID does not exist" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207136 4907 scope.go:117] "RemoveContainer" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" Mar 13 15:22:02 crc kubenswrapper[4907]: E0313 15:22:02.207309 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a\": container with ID starting with 7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a not found: ID does not exist" containerID="7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207329 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a"} err="failed to get container status \"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a\": rpc error: code = NotFound desc = could not find container \"7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a\": container with ID starting with 7f4c959d1547bf070df1ec07daaac52181e0605870a8c576d5c25db49214bf5a not found: ID does not exist" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207357 4907 scope.go:117] "RemoveContainer" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" Mar 13 15:22:02 crc kubenswrapper[4907]: E0313 15:22:02.207570 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948\": container with ID starting with 52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948 not found: ID does not exist" containerID="52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948" Mar 13 15:22:02 crc kubenswrapper[4907]: I0313 15:22:02.207588 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948"} err="failed to get container status \"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948\": rpc error: code = NotFound desc = could not find container \"52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948\": container with ID starting with 52eff7ddefc909e7c1ea25af560e2b9a36da156db25cc2701e4a6f34d753a948 not found: ID does not exist" Mar 13 15:22:03 crc kubenswrapper[4907]: I0313 15:22:03.151064 4907 generic.go:334] "Generic (PLEG): container finished" podID="31cb9010-2aab-475d-aab3-9aa35a538581" containerID="2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519" exitCode=0 Mar 13 15:22:03 crc kubenswrapper[4907]: I0313 15:22:03.151192 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerDied","Data":"2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519"} Mar 13 15:22:03 crc kubenswrapper[4907]: E0313 15:22:03.599595 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:03 crc kubenswrapper[4907]: I0313 15:22:03.791056 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" path="/var/lib/kubelet/pods/8ce4bcd2-1af1-43b6-8e59-d470e9089686/volumes" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.466382 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.629354 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") pod \"31cb9010-2aab-475d-aab3-9aa35a538581\" (UID: \"31cb9010-2aab-475d-aab3-9aa35a538581\") " Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.635076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8" (OuterVolumeSpecName: "kube-api-access-9kjk8") pod "31cb9010-2aab-475d-aab3-9aa35a538581" (UID: "31cb9010-2aab-475d-aab3-9aa35a538581"). InnerVolumeSpecName "kube-api-access-9kjk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.731422 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kjk8\" (UniqueName: \"kubernetes.io/projected/31cb9010-2aab-475d-aab3-9aa35a538581-kube-api-access-9kjk8\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.878241 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:22:04 crc kubenswrapper[4907]: I0313 15:22:04.886806 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556916-gpp98"] Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.166114 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" event={"ID":"31cb9010-2aab-475d-aab3-9aa35a538581","Type":"ContainerDied","Data":"0e6b3e4cee3836f04db13b82a09d26c6bd36e7d4ef4e9ebe75b7a51fdd3a4b97"} Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.166153 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e6b3e4cee3836f04db13b82a09d26c6bd36e7d4ef4e9ebe75b7a51fdd3a4b97" Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.166186 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556922-b7qtk" Mar 13 15:22:05 crc kubenswrapper[4907]: I0313 15:22:05.791392 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5c630da-de73-4f29-89f0-43c5497dd89a" path="/var/lib/kubelet/pods/c5c630da-de73-4f29-89f0-43c5497dd89a/volumes" Mar 13 15:22:13 crc kubenswrapper[4907]: E0313 15:22:13.761836 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:14 crc kubenswrapper[4907]: I0313 15:22:14.782962 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:14 crc kubenswrapper[4907]: E0313 15:22:14.783476 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.672191 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.680209 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-wds2z"] Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.792116 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b501aea0-66d1-4655-b67f-c6a3bacb7cdb" path="/var/lib/kubelet/pods/b501aea0-66d1-4655-b67f-c6a3bacb7cdb/volumes" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.821743 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822081 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" containerName="oc" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822096 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" containerName="oc" Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822111 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822120 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822134 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-content" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822141 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-content" Mar 13 15:22:21 crc kubenswrapper[4907]: E0313 15:22:21.822149 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-utilities" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822157 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="extract-utilities" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822339 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" containerName="oc" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822353 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce4bcd2-1af1-43b6-8e59-d470e9089686" containerName="registry-server" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.822896 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.827796 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.827931 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.828135 4907 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2d28j" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.828753 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.832173 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.981754 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.981805 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:21 crc kubenswrapper[4907]: I0313 15:22:21.981853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.083437 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.083495 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.083541 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.084624 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.084800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.103191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"crc-storage-crc-g69ht\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.143569 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:22 crc kubenswrapper[4907]: I0313 15:22:22.624869 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:23 crc kubenswrapper[4907]: I0313 15:22:23.289756 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g69ht" event={"ID":"9dbd65f5-4738-4d22-a8da-5be0af6211fa","Type":"ContainerStarted","Data":"71e8f6f74869e988544440e1e55a81639417cc910f590802c295e1f053265c1e"} Mar 13 15:22:23 crc kubenswrapper[4907]: E0313 15:22:23.940611 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:24 crc kubenswrapper[4907]: I0313 15:22:24.303224 4907 generic.go:334] "Generic (PLEG): container finished" podID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerID="d45e59416766d20c4f95b995286178dd988b4b6257768b1fef8c82dcc281cd6c" exitCode=0 Mar 13 15:22:24 crc kubenswrapper[4907]: I0313 15:22:24.303300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g69ht" event={"ID":"9dbd65f5-4738-4d22-a8da-5be0af6211fa","Type":"ContainerDied","Data":"d45e59416766d20c4f95b995286178dd988b4b6257768b1fef8c82dcc281cd6c"} Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.650746 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838268 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") pod \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838459 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") pod \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838489 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") pod \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\" (UID: \"9dbd65f5-4738-4d22-a8da-5be0af6211fa\") " Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.838819 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "9dbd65f5-4738-4d22-a8da-5be0af6211fa" (UID: "9dbd65f5-4738-4d22-a8da-5be0af6211fa"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.844732 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg" (OuterVolumeSpecName: "kube-api-access-km7dg") pod "9dbd65f5-4738-4d22-a8da-5be0af6211fa" (UID: "9dbd65f5-4738-4d22-a8da-5be0af6211fa"). InnerVolumeSpecName "kube-api-access-km7dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.877671 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "9dbd65f5-4738-4d22-a8da-5be0af6211fa" (UID: "9dbd65f5-4738-4d22-a8da-5be0af6211fa"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.939896 4907 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/9dbd65f5-4738-4d22-a8da-5be0af6211fa-node-mnt\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.939926 4907 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/9dbd65f5-4738-4d22-a8da-5be0af6211fa-crc-storage\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:25 crc kubenswrapper[4907]: I0313 15:22:25.939936 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km7dg\" (UniqueName: \"kubernetes.io/projected/9dbd65f5-4738-4d22-a8da-5be0af6211fa-kube-api-access-km7dg\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:26 crc kubenswrapper[4907]: I0313 15:22:26.318541 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-g69ht" event={"ID":"9dbd65f5-4738-4d22-a8da-5be0af6211fa","Type":"ContainerDied","Data":"71e8f6f74869e988544440e1e55a81639417cc910f590802c295e1f053265c1e"} Mar 13 15:22:26 crc kubenswrapper[4907]: I0313 15:22:26.318582 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71e8f6f74869e988544440e1e55a81639417cc910f590802c295e1f053265c1e" Mar 13 15:22:26 crc kubenswrapper[4907]: I0313 15:22:26.318608 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-g69ht" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.083213 4907 scope.go:117] "RemoveContainer" containerID="311a031e1ff4ef3b74d10161b82c176889b7b82821fc88f2a4edeca5e5657ee5" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.106503 4907 scope.go:117] "RemoveContainer" containerID="348c87a146da948df0e4139184e97c7692cc71aad3ec14fe9f015dc10660b3b8" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.782081 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:27 crc kubenswrapper[4907]: E0313 15:22:27.782435 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.986206 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:27 crc kubenswrapper[4907]: I0313 15:22:27.993156 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-g69ht"] Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.140830 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-vl976"] Mar 13 15:22:28 crc kubenswrapper[4907]: E0313 15:22:28.141225 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerName="storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.141242 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerName="storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.141405 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" containerName="storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.141971 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.145785 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.146096 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.147073 4907 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-2d28j" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.147134 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.152641 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-vl976"] Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.274137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.274212 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.274248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.375972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.376044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.376074 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.376631 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.379752 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.397932 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"crc-storage-crc-vl976\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.459864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:28 crc kubenswrapper[4907]: I0313 15:22:28.879788 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-vl976"] Mar 13 15:22:29 crc kubenswrapper[4907]: I0313 15:22:29.342128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vl976" event={"ID":"b7d369a5-227b-42bb-9e19-e2cc88a42a24","Type":"ContainerStarted","Data":"762a2cc05f7d23fe18e4297120fcdfa9ad522d3b313575d0337e171ffdddeab6"} Mar 13 15:22:29 crc kubenswrapper[4907]: I0313 15:22:29.791964 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dbd65f5-4738-4d22-a8da-5be0af6211fa" path="/var/lib/kubelet/pods/9dbd65f5-4738-4d22-a8da-5be0af6211fa/volumes" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.281405 4907 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/a6/a6f5c002ec832d6e6a25c08e60f361308d0e752c78eebf8774d041416f5e0be9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260313%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260313T152229Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=a401e230d4c2c642de4064ed340803135dac3ca9bbe99019dcef287f4718d3f1®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=bash&akamai_signature=exp=1773416249~hmac=84d0e01e00b0a4a6dec744898931c695f38a70686f6235dc6c63b84df4e4ef32\": remote error: tls: internal error" image="quay.io/openstack-k8s-operators/bash:latest" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.281600 4907 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:storage,Image:quay.io/openstack-k8s-operators/bash:latest,Command:[bash],Args:[/usr/local/bin/crc-storage.sh],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:PV_NUM,Value:12,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:crc-storage,ReadOnly:true,MountPath:/usr/local/bin/crc-storage.sh,SubPath:create-storage.sh,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:node-mnt,ReadOnly:false,MountPath:/mnt/nodeMnt,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zl6d9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-storage-crc-vl976_crc-storage(b7d369a5-227b-42bb-9e19-e2cc88a42a24): ErrImagePull: parsing image configuration: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/a6/a6f5c002ec832d6e6a25c08e60f361308d0e752c78eebf8774d041416f5e0be9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260313%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260313T152229Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=a401e230d4c2c642de4064ed340803135dac3ca9bbe99019dcef287f4718d3f1®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=bash&akamai_signature=exp=1773416249~hmac=84d0e01e00b0a4a6dec744898931c695f38a70686f6235dc6c63b84df4e4ef32\": remote error: tls: internal error" logger="UnhandledError" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.283042 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"storage\" with ErrImagePull: \"parsing image configuration: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/a6/a6f5c002ec832d6e6a25c08e60f361308d0e752c78eebf8774d041416f5e0be9?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20260313%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20260313T152229Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=a401e230d4c2c642de4064ed340803135dac3ca9bbe99019dcef287f4718d3f1®ion=us-east-1&namespace=openstack-k8s-operators&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=bash&akamai_signature=exp=1773416249~hmac=84d0e01e00b0a4a6dec744898931c695f38a70686f6235dc6c63b84df4e4ef32\\\": remote error: tls: internal error\"" pod="crc-storage/crc-storage-crc-vl976" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" Mar 13 15:22:30 crc kubenswrapper[4907]: E0313 15:22:30.351807 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"storage\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/bash:latest\\\"\"" pod="crc-storage/crc-storage-crc-vl976" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" Mar 13 15:22:34 crc kubenswrapper[4907]: E0313 15:22:34.112033 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:39 crc kubenswrapper[4907]: I0313 15:22:39.782842 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:39 crc kubenswrapper[4907]: E0313 15:22:39.783574 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:44 crc kubenswrapper[4907]: E0313 15:22:44.304861 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:22:46 crc kubenswrapper[4907]: I0313 15:22:46.454331 4907 generic.go:334] "Generic (PLEG): container finished" podID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerID="720dafffc2f4de439a0af5f54cec3d94c4aabe1ca444e2ead9104c7711dcd4fe" exitCode=0 Mar 13 15:22:46 crc kubenswrapper[4907]: I0313 15:22:46.454422 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vl976" event={"ID":"b7d369a5-227b-42bb-9e19-e2cc88a42a24","Type":"ContainerDied","Data":"720dafffc2f4de439a0af5f54cec3d94c4aabe1ca444e2ead9104c7711dcd4fe"} Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.814312 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952304 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") pod \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952390 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") pod \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") pod \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\" (UID: \"b7d369a5-227b-42bb-9e19-e2cc88a42a24\") " Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.952649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "b7d369a5-227b-42bb-9e19-e2cc88a42a24" (UID: "b7d369a5-227b-42bb-9e19-e2cc88a42a24"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.957641 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9" (OuterVolumeSpecName: "kube-api-access-zl6d9") pod "b7d369a5-227b-42bb-9e19-e2cc88a42a24" (UID: "b7d369a5-227b-42bb-9e19-e2cc88a42a24"). InnerVolumeSpecName "kube-api-access-zl6d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:22:47 crc kubenswrapper[4907]: I0313 15:22:47.969902 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "b7d369a5-227b-42bb-9e19-e2cc88a42a24" (UID: "b7d369a5-227b-42bb-9e19-e2cc88a42a24"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.053785 4907 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/b7d369a5-227b-42bb-9e19-e2cc88a42a24-crc-storage\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.053832 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl6d9\" (UniqueName: \"kubernetes.io/projected/b7d369a5-227b-42bb-9e19-e2cc88a42a24-kube-api-access-zl6d9\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.053848 4907 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/b7d369a5-227b-42bb-9e19-e2cc88a42a24-node-mnt\") on node \"crc\" DevicePath \"\"" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.473108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-vl976" event={"ID":"b7d369a5-227b-42bb-9e19-e2cc88a42a24","Type":"ContainerDied","Data":"762a2cc05f7d23fe18e4297120fcdfa9ad522d3b313575d0337e171ffdddeab6"} Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.473146 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-vl976" Mar 13 15:22:48 crc kubenswrapper[4907]: I0313 15:22:48.473156 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="762a2cc05f7d23fe18e4297120fcdfa9ad522d3b313575d0337e171ffdddeab6" Mar 13 15:22:53 crc kubenswrapper[4907]: I0313 15:22:53.782916 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:22:53 crc kubenswrapper[4907]: E0313 15:22:53.783943 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:22:54 crc kubenswrapper[4907]: E0313 15:22:54.463821 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ce4bcd2_1af1_43b6_8e59_d470e9089686.slice/crio-6986ac91f0a1405d74e10dd104ecf7a3a66af918e31112a29fb3477814a7cb12\": RecentStats: unable to find data in memory cache]" Mar 13 15:23:01 crc kubenswrapper[4907]: E0313 15:23:01.811926 4907 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/25607d4d87b5b5a88585c2cd1d2ea551307e53f784e3d597bd98d5b6a061095c/diff" to get inode usage: stat /var/lib/containers/storage/overlay/25607d4d87b5b5a88585c2cd1d2ea551307e53f784e3d597bd98d5b6a061095c/diff: no such file or directory, extraDiskErr: Mar 13 15:23:06 crc kubenswrapper[4907]: I0313 15:23:06.782670 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:23:06 crc kubenswrapper[4907]: E0313 15:23:06.783761 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:23:21 crc kubenswrapper[4907]: I0313 15:23:21.791721 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:23:22 crc kubenswrapper[4907]: I0313 15:23:22.719756 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd"} Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.141937 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:24:00 crc kubenswrapper[4907]: E0313 15:24:00.142854 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerName="storage" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.142871 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerName="storage" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.143089 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7d369a5-227b-42bb-9e19-e2cc88a42a24" containerName="storage" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.143662 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.145447 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.145663 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.149487 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.162517 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.291294 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"auto-csr-approver-29556924-8z5rs\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.393062 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"auto-csr-approver-29556924-8z5rs\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.410541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"auto-csr-approver-29556924-8z5rs\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.467164 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.874336 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:24:00 crc kubenswrapper[4907]: I0313 15:24:00.970822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" event={"ID":"11a50278-3da1-4a63-ac7e-1fcb4ae77f52","Type":"ContainerStarted","Data":"1ce5fdadd767eafa59d4051e216d3c016e27cf39dffbc0f38e628d186e7e2aef"} Mar 13 15:24:02 crc kubenswrapper[4907]: I0313 15:24:02.988404 4907 generic.go:334] "Generic (PLEG): container finished" podID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerID="16a638d449eaf9553071c678dcf92f038cd2ddde8f659099ade22f5e78a0fcae" exitCode=0 Mar 13 15:24:02 crc kubenswrapper[4907]: I0313 15:24:02.988475 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" event={"ID":"11a50278-3da1-4a63-ac7e-1fcb4ae77f52","Type":"ContainerDied","Data":"16a638d449eaf9553071c678dcf92f038cd2ddde8f659099ade22f5e78a0fcae"} Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.350423 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.449384 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") pod \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\" (UID: \"11a50278-3da1-4a63-ac7e-1fcb4ae77f52\") " Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.454947 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7" (OuterVolumeSpecName: "kube-api-access-p2pm7") pod "11a50278-3da1-4a63-ac7e-1fcb4ae77f52" (UID: "11a50278-3da1-4a63-ac7e-1fcb4ae77f52"). InnerVolumeSpecName "kube-api-access-p2pm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:24:04 crc kubenswrapper[4907]: I0313 15:24:04.550406 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2pm7\" (UniqueName: \"kubernetes.io/projected/11a50278-3da1-4a63-ac7e-1fcb4ae77f52-kube-api-access-p2pm7\") on node \"crc\" DevicePath \"\"" Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.004558 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" event={"ID":"11a50278-3da1-4a63-ac7e-1fcb4ae77f52","Type":"ContainerDied","Data":"1ce5fdadd767eafa59d4051e216d3c016e27cf39dffbc0f38e628d186e7e2aef"} Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.005006 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ce5fdadd767eafa59d4051e216d3c016e27cf39dffbc0f38e628d186e7e2aef" Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.004602 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556924-8z5rs" Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.416348 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.422081 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556918-cs4dh"] Mar 13 15:24:05 crc kubenswrapper[4907]: I0313 15:24:05.791334 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdd030a7-3ebb-4fbf-91f2-2ab96e526252" path="/var/lib/kubelet/pods/fdd030a7-3ebb-4fbf-91f2-2ab96e526252/volumes" Mar 13 15:24:27 crc kubenswrapper[4907]: I0313 15:24:27.207304 4907 scope.go:117] "RemoveContainer" containerID="e984813306df1ab82de9a9dd04affc9e989fc2092eb575a81e8fc5794b033f1a" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.610047 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:36 crc kubenswrapper[4907]: E0313 15:25:36.611011 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerName="oc" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.611028 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerName="oc" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.611186 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" containerName="oc" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.612392 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.622806 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.766485 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.766728 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.766853 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.867952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868105 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.868901 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.891012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"community-operators-skd22\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:36 crc kubenswrapper[4907]: I0313 15:25:36.933457 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.449297 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.656586 4907 generic.go:334] "Generic (PLEG): container finished" podID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" exitCode=0 Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.656629 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d"} Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.656654 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerStarted","Data":"4ff9f5b55970d7e99e658759b98269c42d65a135c1963e0a4852afb8d683eb2f"} Mar 13 15:25:37 crc kubenswrapper[4907]: I0313 15:25:37.659711 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:25:39 crc kubenswrapper[4907]: I0313 15:25:39.677794 4907 generic.go:334] "Generic (PLEG): container finished" podID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" exitCode=0 Mar 13 15:25:39 crc kubenswrapper[4907]: I0313 15:25:39.677850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa"} Mar 13 15:25:40 crc kubenswrapper[4907]: I0313 15:25:40.688502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerStarted","Data":"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f"} Mar 13 15:25:46 crc kubenswrapper[4907]: I0313 15:25:46.933951 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:46 crc kubenswrapper[4907]: I0313 15:25:46.934524 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:46 crc kubenswrapper[4907]: I0313 15:25:46.985258 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:47 crc kubenswrapper[4907]: I0313 15:25:47.008404 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-skd22" podStartSLOduration=8.609087121 podStartE2EDuration="11.008377126s" podCreationTimestamp="2026-03-13 15:25:36 +0000 UTC" firstStartedPulling="2026-03-13 15:25:37.659441848 +0000 UTC m=+4836.559229537" lastFinishedPulling="2026-03-13 15:25:40.058731853 +0000 UTC m=+4838.958519542" observedRunningTime="2026-03-13 15:25:40.718072509 +0000 UTC m=+4839.617860208" watchObservedRunningTime="2026-03-13 15:25:47.008377126 +0000 UTC m=+4845.908164825" Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.041124 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.041323 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.041375 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:25:48 crc kubenswrapper[4907]: I0313 15:25:48.085672 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:49 crc kubenswrapper[4907]: I0313 15:25:49.743860 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-skd22" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" containerID="cri-o://61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" gracePeriod=2 Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.182223 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.372439 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") pod \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.372606 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") pod \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.372638 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") pod \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\" (UID: \"6454399b-51ab-41f1-b4db-bbe12b4b7c95\") " Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.407133 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m" (OuterVolumeSpecName: "kube-api-access-z4r6m") pod "6454399b-51ab-41f1-b4db-bbe12b4b7c95" (UID: "6454399b-51ab-41f1-b4db-bbe12b4b7c95"). InnerVolumeSpecName "kube-api-access-z4r6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.411209 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities" (OuterVolumeSpecName: "utilities") pod "6454399b-51ab-41f1-b4db-bbe12b4b7c95" (UID: "6454399b-51ab-41f1-b4db-bbe12b4b7c95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.451150 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6454399b-51ab-41f1-b4db-bbe12b4b7c95" (UID: "6454399b-51ab-41f1-b4db-bbe12b4b7c95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.473913 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4r6m\" (UniqueName: \"kubernetes.io/projected/6454399b-51ab-41f1-b4db-bbe12b4b7c95-kube-api-access-z4r6m\") on node \"crc\" DevicePath \"\"" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.473946 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.473961 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6454399b-51ab-41f1-b4db-bbe12b4b7c95-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753245 4907 generic.go:334] "Generic (PLEG): container finished" podID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" exitCode=0 Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753295 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f"} Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753319 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-skd22" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753348 4907 scope.go:117] "RemoveContainer" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.753334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-skd22" event={"ID":"6454399b-51ab-41f1-b4db-bbe12b4b7c95","Type":"ContainerDied","Data":"4ff9f5b55970d7e99e658759b98269c42d65a135c1963e0a4852afb8d683eb2f"} Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.773336 4907 scope.go:117] "RemoveContainer" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.794079 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.799416 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-skd22"] Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.808230 4907 scope.go:117] "RemoveContainer" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.824672 4907 scope.go:117] "RemoveContainer" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" Mar 13 15:25:50 crc kubenswrapper[4907]: E0313 15:25:50.825264 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f\": container with ID starting with 61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f not found: ID does not exist" containerID="61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825297 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f"} err="failed to get container status \"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f\": rpc error: code = NotFound desc = could not find container \"61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f\": container with ID starting with 61ac5444abd6e1ac4c4de07ff227d8deea674a73498eeb60cb027d561df5df1f not found: ID does not exist" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825326 4907 scope.go:117] "RemoveContainer" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" Mar 13 15:25:50 crc kubenswrapper[4907]: E0313 15:25:50.825593 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa\": container with ID starting with 40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa not found: ID does not exist" containerID="40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825643 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa"} err="failed to get container status \"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa\": rpc error: code = NotFound desc = could not find container \"40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa\": container with ID starting with 40a57078c576bd33006fc990f6f567d9963ea1da66645391e9fcf035de5e33fa not found: ID does not exist" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.825671 4907 scope.go:117] "RemoveContainer" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" Mar 13 15:25:50 crc kubenswrapper[4907]: E0313 15:25:50.826114 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d\": container with ID starting with c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d not found: ID does not exist" containerID="c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d" Mar 13 15:25:50 crc kubenswrapper[4907]: I0313 15:25:50.826135 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d"} err="failed to get container status \"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d\": rpc error: code = NotFound desc = could not find container \"c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d\": container with ID starting with c1c98d52ad0aba6e93f5f3a8f2f40fa1e06cc70b71f70d32c0e3077df5f68e0d not found: ID does not exist" Mar 13 15:25:51 crc kubenswrapper[4907]: I0313 15:25:51.791638 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" path="/var/lib/kubelet/pods/6454399b-51ab-41f1-b4db-bbe12b4b7c95/volumes" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.139144 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:26:00 crc kubenswrapper[4907]: E0313 15:26:00.139993 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-content" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140014 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-content" Mar 13 15:26:00 crc kubenswrapper[4907]: E0313 15:26:00.140030 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140038 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" Mar 13 15:26:00 crc kubenswrapper[4907]: E0313 15:26:00.140058 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-utilities" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140069 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="extract-utilities" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.140321 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6454399b-51ab-41f1-b4db-bbe12b4b7c95" containerName="registry-server" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.143304 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.145216 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.145242 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.146422 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.154905 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.307527 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"auto-csr-approver-29556926-r9tzc\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.408629 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"auto-csr-approver-29556926-r9tzc\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.430326 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"auto-csr-approver-29556926-r9tzc\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.469411 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:00 crc kubenswrapper[4907]: I0313 15:26:00.876336 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:26:01 crc kubenswrapper[4907]: I0313 15:26:01.824852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" event={"ID":"81a286dc-1af0-43f3-ad06-0e7d42141a0a","Type":"ContainerStarted","Data":"1dee45477813eb103f3973c857556e21148b420ea81f710f1f3bbf6e830f1279"} Mar 13 15:26:02 crc kubenswrapper[4907]: I0313 15:26:02.833301 4907 generic.go:334] "Generic (PLEG): container finished" podID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerID="fdb72e54eff786808e8d908105afbd91c2b9b22c84cb13b6b15ad7f081283322" exitCode=0 Mar 13 15:26:02 crc kubenswrapper[4907]: I0313 15:26:02.833387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" event={"ID":"81a286dc-1af0-43f3-ad06-0e7d42141a0a","Type":"ContainerDied","Data":"fdb72e54eff786808e8d908105afbd91c2b9b22c84cb13b6b15ad7f081283322"} Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.083242 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.160217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") pod \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\" (UID: \"81a286dc-1af0-43f3-ad06-0e7d42141a0a\") " Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.168123 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns" (OuterVolumeSpecName: "kube-api-access-nntns") pod "81a286dc-1af0-43f3-ad06-0e7d42141a0a" (UID: "81a286dc-1af0-43f3-ad06-0e7d42141a0a"). InnerVolumeSpecName "kube-api-access-nntns". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.261755 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nntns\" (UniqueName: \"kubernetes.io/projected/81a286dc-1af0-43f3-ad06-0e7d42141a0a-kube-api-access-nntns\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.847333 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" event={"ID":"81a286dc-1af0-43f3-ad06-0e7d42141a0a","Type":"ContainerDied","Data":"1dee45477813eb103f3973c857556e21148b420ea81f710f1f3bbf6e830f1279"} Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.847715 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dee45477813eb103f3973c857556e21148b420ea81f710f1f3bbf6e830f1279" Mar 13 15:26:04 crc kubenswrapper[4907]: I0313 15:26:04.847372 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556926-r9tzc" Mar 13 15:26:05 crc kubenswrapper[4907]: I0313 15:26:05.154011 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:26:05 crc kubenswrapper[4907]: I0313 15:26:05.160540 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556920-lrh8k"] Mar 13 15:26:05 crc kubenswrapper[4907]: I0313 15:26:05.791252 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925bf3a9-9978-49a9-8ac3-7c92a928adb7" path="/var/lib/kubelet/pods/925bf3a9-9978-49a9-8ac3-7c92a928adb7/volumes" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.923936 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:12 crc kubenswrapper[4907]: E0313 15:26:12.924605 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerName="oc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.924618 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerName="oc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.924733 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" containerName="oc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.925418 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.928520 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.928700 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-5b4k5" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.928820 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.929426 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.929515 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.942648 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.972869 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.972941 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:12 crc kubenswrapper[4907]: I0313 15:26:12.972964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.074210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.074260 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.074287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.075314 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.075519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.092212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"dnsmasq-dns-7fc7c884dc-98jxk\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.203027 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.204173 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.214476 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.242381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.377868 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.378145 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.378230 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.479650 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.479842 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.479963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.480933 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.480929 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.500323 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"dnsmasq-dns-7c95686bd5-8kb7m\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.534094 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.692732 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:13 crc kubenswrapper[4907]: W0313 15:26:13.698305 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cdac990_bf65_48a3_8892_b375611f1e70.slice/crio-77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad WatchSource:0}: Error finding container 77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad: Status 404 returned error can't find the container with id 77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.909615 4907 generic.go:334] "Generic (PLEG): container finished" podID="2cdac990-bf65-48a3-8892-b375611f1e70" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" exitCode=0 Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.909665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerDied","Data":"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73"} Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.909729 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerStarted","Data":"77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad"} Mar 13 15:26:13 crc kubenswrapper[4907]: I0313 15:26:13.995719 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:26:14 crc kubenswrapper[4907]: W0313 15:26:14.002098 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1b7f262_ac9a_4a7e_ae94_6ce13d4105a2.slice/crio-18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058 WatchSource:0}: Error finding container 18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058: Status 404 returned error can't find the container with id 18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058 Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.080864 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.081937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.084138 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.084488 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.085373 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vlpfs" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.085613 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.085705 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.095298 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.196790 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.196869 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.196925 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197450 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197625 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197802 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.197850 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299551 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299644 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299670 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299739 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299766 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.299801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.300126 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.300155 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.301012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.301538 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303453 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303484 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c926015d5702d92351724596ba892321797e624e5f7a01b00404b7d1c8a17788/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.303923 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.304230 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.326904 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.337205 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.399367 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.400851 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.402814 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403146 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403276 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403269 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-grjtp" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.403439 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.417047 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.469241 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502808 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502906 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502938 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.502966 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503014 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503089 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.503190 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.604953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605001 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605037 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605061 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605104 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.605178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.606272 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.606524 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.612012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.612025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.612618 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.613620 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.615020 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.615047 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d84a704eae1ca42fe329ab2f9d170d2a410be00de88c5e0a709d7b62965c26cb/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.615107 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.630829 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.643475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.720940 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.893616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.918143 4907 generic.go:334] "Generic (PLEG): container finished" podID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" exitCode=0 Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.918214 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerDied","Data":"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.918245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerStarted","Data":"18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.921749 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerStarted","Data":"7fa7486d25e0b1f49302f99040c36181735134e64166f989efe3f3ae26956d78"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.926316 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerStarted","Data":"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8"} Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.928635 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:14 crc kubenswrapper[4907]: I0313 15:26:14.965228 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" podStartSLOduration=2.9652065050000003 podStartE2EDuration="2.965206505s" podCreationTimestamp="2026-03-13 15:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:14.963069347 +0000 UTC m=+4873.862857036" watchObservedRunningTime="2026-03-13 15:26:14.965206505 +0000 UTC m=+4873.864994194" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.124863 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: W0313 15:26:15.129411 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75f775f2_b77c_4ff4_80e4_12f6d3e2d0fd.slice/crio-24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788 WatchSource:0}: Error finding container 24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788: Status 404 returned error can't find the container with id 24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788 Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.484549 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.490706 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.498409 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.498762 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xl4lf" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.499079 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.499316 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.502801 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.508849 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618656 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618705 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-config-data-default\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618737 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/585e3be7-1e17-443c-b430-0c03727167cb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618802 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvvpz\" (UniqueName: \"kubernetes.io/projected/585e3be7-1e17-443c-b430-0c03727167cb-kube-api-access-rvvpz\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618844 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-kolla-config\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618897 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.618963 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720651 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720734 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-config-data-default\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720775 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/585e3be7-1e17-443c-b430-0c03727167cb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvvpz\" (UniqueName: \"kubernetes.io/projected/585e3be7-1e17-443c-b430-0c03727167cb-kube-api-access-rvvpz\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720827 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-kolla-config\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720868 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.720964 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.722000 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/585e3be7-1e17-443c-b430-0c03727167cb-config-data-generated\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.722364 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-operator-scripts\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.722665 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-kolla-config\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.723071 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/585e3be7-1e17-443c-b430-0c03727167cb-config-data-default\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.726183 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.726225 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2835f978c80613ada9b2606f7e7b901545616ef9a453011deca8ef169d3eb6f9/globalmount\"" pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.806861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.807371 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/585e3be7-1e17-443c-b430-0c03727167cb-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.808279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvvpz\" (UniqueName: \"kubernetes.io/projected/585e3be7-1e17-443c-b430-0c03727167cb-kube-api-access-rvvpz\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.933415 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.934456 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.937418 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.937707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerStarted","Data":"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea"} Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.937754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.939165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerStarted","Data":"24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788"} Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.943751 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-s66hx" Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.946119 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 15:26:15 crc kubenswrapper[4907]: I0313 15:26:15.979712 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" podStartSLOduration=2.979689441 podStartE2EDuration="2.979689441s" podCreationTimestamp="2026-03-13 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:15.973397779 +0000 UTC m=+4874.873185468" watchObservedRunningTime="2026-03-13 15:26:15.979689441 +0000 UTC m=+4874.879477140" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.025041 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-config-data\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.025355 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpmrh\" (UniqueName: \"kubernetes.io/projected/8c712e4f-4999-476b-970f-eae3189b4e59-kube-api-access-qpmrh\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.025418 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-kolla-config\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.121522 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2e0cba70-4127-4af3-a5ff-dca1a2fa2510\") pod \"openstack-galera-0\" (UID: \"585e3be7-1e17-443c-b430-0c03727167cb\") " pod="openstack/openstack-galera-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.126429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-config-data\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.126508 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpmrh\" (UniqueName: \"kubernetes.io/projected/8c712e4f-4999-476b-970f-eae3189b4e59-kube-api-access-qpmrh\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.126565 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-kolla-config\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.127407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-kolla-config\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.127438 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c712e4f-4999-476b-970f-eae3189b4e59-config-data\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.161525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpmrh\" (UniqueName: \"kubernetes.io/projected/8c712e4f-4999-476b-970f-eae3189b4e59-kube-api-access-qpmrh\") pod \"memcached-0\" (UID: \"8c712e4f-4999-476b-970f-eae3189b4e59\") " pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.250360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.412261 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.724546 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.840975 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 15:26:16 crc kubenswrapper[4907]: W0313 15:26:16.913177 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c712e4f_4999_476b_970f_eae3189b4e59.slice/crio-a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02 WatchSource:0}: Error finding container a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02: Status 404 returned error can't find the container with id a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02 Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.948391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerStarted","Data":"bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272"} Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.950047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8c712e4f-4999-476b-970f-eae3189b4e59","Type":"ContainerStarted","Data":"a0f9a382e21caec5c6f87f1cc77b784feb6313013846ee54439ff8d42725be02"} Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.951068 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerStarted","Data":"5c54f8dc7a3d2ec4cd9dcf35e8b0c95c41f32bfe48a3f1eab285bc0ae0bac5eb"} Mar 13 15:26:16 crc kubenswrapper[4907]: I0313 15:26:16.952775 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerStarted","Data":"0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.101445 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.102937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.106258 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.106595 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.109695 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-stnw8" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.115066 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.117461 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143480 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143570 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143688 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143822 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjp8c\" (UniqueName: \"kubernetes.io/projected/edb521bb-211e-48c0-b644-8dc44b235feb-kube-api-access-pjp8c\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143915 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.143973 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.144047 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245545 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245638 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjp8c\" (UniqueName: \"kubernetes.io/projected/edb521bb-211e-48c0-b644-8dc44b235feb-kube-api-access-pjp8c\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245675 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245699 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245731 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245804 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.245836 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.246086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/edb521bb-211e-48c0-b644-8dc44b235feb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.246304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.247109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/edb521bb-211e-48c0-b644-8dc44b235feb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.249825 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.249934 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/edb521bb-211e-48c0-b644-8dc44b235feb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.251507 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.251537 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bc929da5e54fcbc45e8b64e6c7cf70de64445330602309f6aca5ea888f5267cb/globalmount\"" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.263652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjp8c\" (UniqueName: \"kubernetes.io/projected/edb521bb-211e-48c0-b644-8dc44b235feb-kube-api-access-pjp8c\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.276480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdc965c-6ae3-457e-9054-e9bd61be219a\") pod \"openstack-cell1-galera-0\" (UID: \"edb521bb-211e-48c0-b644-8dc44b235feb\") " pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.383129 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.828629 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 15:26:17 crc kubenswrapper[4907]: W0313 15:26:17.839383 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedb521bb_211e_48c0_b644_8dc44b235feb.slice/crio-0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061 WatchSource:0}: Error finding container 0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061: Status 404 returned error can't find the container with id 0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061 Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.962376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerStarted","Data":"0055e47f11b76f561cdcf0890df26a527b6eb47e8ee371daa37c8c6e62170061"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.963933 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8c712e4f-4999-476b-970f-eae3189b4e59","Type":"ContainerStarted","Data":"72775b2f66df067aaded0627229fbd6c0b8d5f1a8301a47432b9b81d596998f8"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.964817 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.967781 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerStarted","Data":"f3b5304b835c93b9e62b5aa23ff48efa75463f10277227bd0d73800bbbe03088"} Mar 13 15:26:17 crc kubenswrapper[4907]: I0313 15:26:17.983662 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.983596097 podStartE2EDuration="2.983596097s" podCreationTimestamp="2026-03-13 15:26:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:17.980145642 +0000 UTC m=+4876.879933341" watchObservedRunningTime="2026-03-13 15:26:17.983596097 +0000 UTC m=+4876.883383776" Mar 13 15:26:18 crc kubenswrapper[4907]: I0313 15:26:18.041785 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:26:18 crc kubenswrapper[4907]: I0313 15:26:18.041851 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:26:18 crc kubenswrapper[4907]: I0313 15:26:18.979640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerStarted","Data":"0a58e80cfc5f151adf7f48c9420de170b24a2f153322cc550036bc388dca8316"} Mar 13 15:26:20 crc kubenswrapper[4907]: I0313 15:26:20.997150 4907 generic.go:334] "Generic (PLEG): container finished" podID="585e3be7-1e17-443c-b430-0c03727167cb" containerID="f3b5304b835c93b9e62b5aa23ff48efa75463f10277227bd0d73800bbbe03088" exitCode=0 Mar 13 15:26:20 crc kubenswrapper[4907]: I0313 15:26:20.997264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerDied","Data":"f3b5304b835c93b9e62b5aa23ff48efa75463f10277227bd0d73800bbbe03088"} Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.009272 4907 generic.go:334] "Generic (PLEG): container finished" podID="edb521bb-211e-48c0-b644-8dc44b235feb" containerID="0a58e80cfc5f151adf7f48c9420de170b24a2f153322cc550036bc388dca8316" exitCode=0 Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.009382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerDied","Data":"0a58e80cfc5f151adf7f48c9420de170b24a2f153322cc550036bc388dca8316"} Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.012684 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"585e3be7-1e17-443c-b430-0c03727167cb","Type":"ContainerStarted","Data":"8d2b4da29a65b464b3858d352aa2641c49967484e6c693a9d41c419714ea3020"} Mar 13 15:26:22 crc kubenswrapper[4907]: I0313 15:26:22.071464 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.071443857 podStartE2EDuration="8.071443857s" podCreationTimestamp="2026-03-13 15:26:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:22.061659299 +0000 UTC m=+4880.961446988" watchObservedRunningTime="2026-03-13 15:26:22.071443857 +0000 UTC m=+4880.971231546" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.020991 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"edb521bb-211e-48c0-b644-8dc44b235feb","Type":"ContainerStarted","Data":"775073a54525f805cb5e13eaea48657add5873f67ca298d457a515b2d40981c7"} Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.049748 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.049726463 podStartE2EDuration="7.049726463s" podCreationTimestamp="2026-03-13 15:26:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:23.044343436 +0000 UTC m=+4881.944131145" watchObservedRunningTime="2026-03-13 15:26:23.049726463 +0000 UTC m=+4881.949514172" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.244167 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.535494 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:26:23 crc kubenswrapper[4907]: I0313 15:26:23.586971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.026641 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" containerID="cri-o://4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" gracePeriod=10 Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.426703 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.559158 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") pod \"2cdac990-bf65-48a3-8892-b375611f1e70\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.559218 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") pod \"2cdac990-bf65-48a3-8892-b375611f1e70\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.559266 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") pod \"2cdac990-bf65-48a3-8892-b375611f1e70\" (UID: \"2cdac990-bf65-48a3-8892-b375611f1e70\") " Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.564700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9" (OuterVolumeSpecName: "kube-api-access-6c4r9") pod "2cdac990-bf65-48a3-8892-b375611f1e70" (UID: "2cdac990-bf65-48a3-8892-b375611f1e70"). InnerVolumeSpecName "kube-api-access-6c4r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.599660 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2cdac990-bf65-48a3-8892-b375611f1e70" (UID: "2cdac990-bf65-48a3-8892-b375611f1e70"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.599844 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config" (OuterVolumeSpecName: "config") pod "2cdac990-bf65-48a3-8892-b375611f1e70" (UID: "2cdac990-bf65-48a3-8892-b375611f1e70"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.660659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c4r9\" (UniqueName: \"kubernetes.io/projected/2cdac990-bf65-48a3-8892-b375611f1e70-kube-api-access-6c4r9\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.660869 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:24 crc kubenswrapper[4907]: I0313 15:26:24.660990 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cdac990-bf65-48a3-8892-b375611f1e70-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035175 4907 generic.go:334] "Generic (PLEG): container finished" podID="2cdac990-bf65-48a3-8892-b375611f1e70" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" exitCode=0 Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035259 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerDied","Data":"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8"} Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035460 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fc7c884dc-98jxk" event={"ID":"2cdac990-bf65-48a3-8892-b375611f1e70","Type":"ContainerDied","Data":"77a29a36e6edcd7453504c99bd8d6185ca0a664cb336729564e5ea2647a574ad"} Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.035494 4907 scope.go:117] "RemoveContainer" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.079136 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.083915 4907 scope.go:117] "RemoveContainer" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.085239 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fc7c884dc-98jxk"] Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101061 4907 scope.go:117] "RemoveContainer" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" Mar 13 15:26:25 crc kubenswrapper[4907]: E0313 15:26:25.101539 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8\": container with ID starting with 4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8 not found: ID does not exist" containerID="4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101577 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8"} err="failed to get container status \"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8\": rpc error: code = NotFound desc = could not find container \"4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8\": container with ID starting with 4d2e0d92c2b302461c3a8b6d60213e7b81411578e2bc548ac86409412c045bf8 not found: ID does not exist" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101603 4907 scope.go:117] "RemoveContainer" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" Mar 13 15:26:25 crc kubenswrapper[4907]: E0313 15:26:25.101895 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73\": container with ID starting with 380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73 not found: ID does not exist" containerID="380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.101922 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73"} err="failed to get container status \"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73\": rpc error: code = NotFound desc = could not find container \"380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73\": container with ID starting with 380d0c10f65356a76638952335ee06865a37a23f94d9e43041a4522fb0545d73 not found: ID does not exist" Mar 13 15:26:25 crc kubenswrapper[4907]: I0313 15:26:25.791940 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" path="/var/lib/kubelet/pods/2cdac990-bf65-48a3-8892-b375611f1e70/volumes" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.252360 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.413218 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.413273 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 15:26:26 crc kubenswrapper[4907]: I0313 15:26:26.488754 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 15:26:26 crc kubenswrapper[4907]: E0313 15:26:26.643016 4907 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.174:56862->38.102.83.174:33349: write tcp 38.102.83.174:56862->38.102.83.174:33349: write: connection reset by peer Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.122622 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.289929 4907 scope.go:117] "RemoveContainer" containerID="e5083f5e7cb4434ec8bff6d3c9df8bdd433195c0fc9dbe495454864580276a9b" Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.384149 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:27 crc kubenswrapper[4907]: I0313 15:26:27.384210 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:29 crc kubenswrapper[4907]: I0313 15:26:29.675068 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:29 crc kubenswrapper[4907]: I0313 15:26:29.764744 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.448591 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:34 crc kubenswrapper[4907]: E0313 15:26:34.449486 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="init" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.449504 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="init" Mar 13 15:26:34 crc kubenswrapper[4907]: E0313 15:26:34.449524 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.449532 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.449742 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cdac990-bf65-48a3-8892-b375611f1e70" containerName="dnsmasq-dns" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.450344 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.453312 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.455841 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.519803 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.519968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.621680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.621851 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.622720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.643246 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"root-account-create-update-2gzz9\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:34 crc kubenswrapper[4907]: I0313 15:26:34.770497 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:35 crc kubenswrapper[4907]: I0313 15:26:35.184226 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:35 crc kubenswrapper[4907]: W0313 15:26:35.195368 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0ad87e8_d3b1_4f00_8f3d_b3b95e9be127.slice/crio-ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b WatchSource:0}: Error finding container ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b: Status 404 returned error can't find the container with id ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b Mar 13 15:26:36 crc kubenswrapper[4907]: I0313 15:26:36.118932 4907 generic.go:334] "Generic (PLEG): container finished" podID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerID="8f8264e04faa67742d13e20b2a8a6a04d3b6912538ef54a5202a7e841e98d29e" exitCode=0 Mar 13 15:26:36 crc kubenswrapper[4907]: I0313 15:26:36.119067 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2gzz9" event={"ID":"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127","Type":"ContainerDied","Data":"8f8264e04faa67742d13e20b2a8a6a04d3b6912538ef54a5202a7e841e98d29e"} Mar 13 15:26:36 crc kubenswrapper[4907]: I0313 15:26:36.119263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2gzz9" event={"ID":"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127","Type":"ContainerStarted","Data":"ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b"} Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.389310 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.560269 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") pod \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.560446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") pod \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\" (UID: \"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127\") " Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.561108 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" (UID: "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.571754 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb" (OuterVolumeSpecName: "kube-api-access-t52tb") pod "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" (UID: "b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127"). InnerVolumeSpecName "kube-api-access-t52tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.662499 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t52tb\" (UniqueName: \"kubernetes.io/projected/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-kube-api-access-t52tb\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:37 crc kubenswrapper[4907]: I0313 15:26:37.662532 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:38 crc kubenswrapper[4907]: I0313 15:26:38.138513 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-2gzz9" event={"ID":"b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127","Type":"ContainerDied","Data":"ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b"} Mar 13 15:26:38 crc kubenswrapper[4907]: I0313 15:26:38.138739 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea9cce4ceae5a244b75b04140ee1e964bc68e196b8c77a4abefc008e564e1a4b" Mar 13 15:26:38 crc kubenswrapper[4907]: I0313 15:26:38.138598 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-2gzz9" Mar 13 15:26:41 crc kubenswrapper[4907]: I0313 15:26:41.094610 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:41 crc kubenswrapper[4907]: I0313 15:26:41.102337 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-2gzz9"] Mar 13 15:26:41 crc kubenswrapper[4907]: I0313 15:26:41.799569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" path="/var/lib/kubelet/pods/b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127/volumes" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.097060 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:26:46 crc kubenswrapper[4907]: E0313 15:26:46.097605 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerName="mariadb-account-create-update" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.097618 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerName="mariadb-account-create-update" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.097803 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ad87e8-d3b1-4f00-8f3d-b3b95e9be127" containerName="mariadb-account-create-update" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.098534 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.101418 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.148191 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.196543 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.196606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.298441 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.298516 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.299428 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.320373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"root-account-create-update-l8vbt\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.431120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:46 crc kubenswrapper[4907]: I0313 15:26:46.863367 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:26:47 crc kubenswrapper[4907]: I0313 15:26:47.240580 4907 generic.go:334] "Generic (PLEG): container finished" podID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerID="bddac8c66ea0090f3f484cd37e970a8a28f550e7e44321363049a8343bf92762" exitCode=0 Mar 13 15:26:47 crc kubenswrapper[4907]: I0313 15:26:47.240715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-l8vbt" event={"ID":"1d5a861d-b4ca-4936-9fac-5aa23fcaa317","Type":"ContainerDied","Data":"bddac8c66ea0090f3f484cd37e970a8a28f550e7e44321363049a8343bf92762"} Mar 13 15:26:47 crc kubenswrapper[4907]: I0313 15:26:47.241065 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-l8vbt" event={"ID":"1d5a861d-b4ca-4936-9fac-5aa23fcaa317","Type":"ContainerStarted","Data":"e1eea7ae7248a157798b49da39f331a2cd401dacad63dfa779a01db360a7bbda"} Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.041851 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.041979 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.042059 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.043224 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.043360 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd" gracePeriod=600 Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.263348 4907 generic.go:334] "Generic (PLEG): container finished" podID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerID="bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272" exitCode=0 Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.263621 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerDied","Data":"bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272"} Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.268238 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd" exitCode=0 Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.268298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd"} Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.268346 4907 scope.go:117] "RemoveContainer" containerID="dd934d7c5fc7e95362f79c2301ff40920ce7e3112722cbc11a053a56c085e05c" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.497644 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.533306 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") pod \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.533466 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") pod \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\" (UID: \"1d5a861d-b4ca-4936-9fac-5aa23fcaa317\") " Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.533849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1d5a861d-b4ca-4936-9fac-5aa23fcaa317" (UID: "1d5a861d-b4ca-4936-9fac-5aa23fcaa317"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.538021 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22" (OuterVolumeSpecName: "kube-api-access-plc22") pod "1d5a861d-b4ca-4936-9fac-5aa23fcaa317" (UID: "1d5a861d-b4ca-4936-9fac-5aa23fcaa317"). InnerVolumeSpecName "kube-api-access-plc22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.635267 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plc22\" (UniqueName: \"kubernetes.io/projected/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-kube-api-access-plc22\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:48 crc kubenswrapper[4907]: I0313 15:26:48.635299 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1d5a861d-b4ca-4936-9fac-5aa23fcaa317-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.280784 4907 generic.go:334] "Generic (PLEG): container finished" podID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerID="0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53" exitCode=0 Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.280933 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerDied","Data":"0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.284376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerStarted","Data":"c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.284583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.288523 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-l8vbt" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.288646 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-l8vbt" event={"ID":"1d5a861d-b4ca-4936-9fac-5aa23fcaa317","Type":"ContainerDied","Data":"e1eea7ae7248a157798b49da39f331a2cd401dacad63dfa779a01db360a7bbda"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.288673 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1eea7ae7248a157798b49da39f331a2cd401dacad63dfa779a01db360a7bbda" Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.292181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596"} Mar 13 15:26:49 crc kubenswrapper[4907]: I0313 15:26:49.371111 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.3710826 podStartE2EDuration="36.3710826s" podCreationTimestamp="2026-03-13 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:49.365643401 +0000 UTC m=+4908.265431090" watchObservedRunningTime="2026-03-13 15:26:49.3710826 +0000 UTC m=+4908.270870299" Mar 13 15:26:50 crc kubenswrapper[4907]: I0313 15:26:50.299899 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerStarted","Data":"83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a"} Mar 13 15:26:50 crc kubenswrapper[4907]: I0313 15:26:50.300355 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:26:50 crc kubenswrapper[4907]: I0313 15:26:50.318923 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.318904363 podStartE2EDuration="37.318904363s" podCreationTimestamp="2026-03-13 15:26:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:26:50.317532926 +0000 UTC m=+4909.217320615" watchObservedRunningTime="2026-03-13 15:26:50.318904363 +0000 UTC m=+4909.218692052" Mar 13 15:27:04 crc kubenswrapper[4907]: I0313 15:27:04.472504 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 15:27:04 crc kubenswrapper[4907]: I0313 15:27:04.723712 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.179088 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:27:10 crc kubenswrapper[4907]: E0313 15:27:10.179957 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerName="mariadb-account-create-update" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.179975 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerName="mariadb-account-create-update" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.180196 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" containerName="mariadb-account-create-update" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.181085 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.198010 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.362832 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.362909 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.363254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.464847 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.465021 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.465053 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.466424 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.466479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.492429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"dnsmasq-dns-684c864bc9-p4j95\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.500192 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:10 crc kubenswrapper[4907]: I0313 15:27:10.940809 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.439747 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.481050 4907 generic.go:334] "Generic (PLEG): container finished" podID="6c367cd2-6ece-429d-a930-c1ec11175248" containerID="dfbeef34d4d9c1c564438050b366f2c5cc11365a8117cb19955310b23bc22e05" exitCode=0 Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.481352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerDied","Data":"dfbeef34d4d9c1c564438050b366f2c5cc11365a8117cb19955310b23bc22e05"} Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.482015 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerStarted","Data":"bc53c9c5cde7280652faad4d880d4ee4a2f7625ff891a79d0f242d66a55b312a"} Mar 13 15:27:11 crc kubenswrapper[4907]: I0313 15:27:11.492740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:12 crc kubenswrapper[4907]: I0313 15:27:12.517854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerStarted","Data":"1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2"} Mar 13 15:27:12 crc kubenswrapper[4907]: I0313 15:27:12.541955 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" podStartSLOduration=2.5419383939999998 podStartE2EDuration="2.541938394s" podCreationTimestamp="2026-03-13 15:27:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:27:12.536272549 +0000 UTC m=+4931.436060258" watchObservedRunningTime="2026-03-13 15:27:12.541938394 +0000 UTC m=+4931.441726083" Mar 13 15:27:13 crc kubenswrapper[4907]: I0313 15:27:13.172982 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" containerID="cri-o://c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36" gracePeriod=604799 Mar 13 15:27:13 crc kubenswrapper[4907]: I0313 15:27:13.185204 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" containerID="cri-o://83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a" gracePeriod=604799 Mar 13 15:27:13 crc kubenswrapper[4907]: I0313 15:27:13.525963 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:14 crc kubenswrapper[4907]: I0313 15:27:14.470905 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.22:5672: connect: connection refused" Mar 13 15:27:14 crc kubenswrapper[4907]: I0313 15:27:14.722199 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.23:5672: connect: connection refused" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.603868 4907 generic.go:334] "Generic (PLEG): container finished" podID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerID="83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a" exitCode=0 Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.603922 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerDied","Data":"83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a"} Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.604462 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd","Type":"ContainerDied","Data":"24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788"} Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.604481 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24618a7621fc23ef2173f3056fdb4f82fc6d9f9738f1b8f67a17813327203788" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.608276 4907 generic.go:334] "Generic (PLEG): container finished" podID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerID="c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36" exitCode=0 Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.608322 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerDied","Data":"c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36"} Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.613819 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726610 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726749 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726815 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726914 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.726991 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") pod \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\" (UID: \"75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.727484 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.728020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.728300 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.734099 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.735017 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info" (OuterVolumeSpecName: "pod-info") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.735103 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j" (OuterVolumeSpecName: "kube-api-access-vhr7j") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "kube-api-access-vhr7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.737604 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f" (OuterVolumeSpecName: "persistence") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "pvc-7c39547e-1168-404d-865b-13cfd1526f6f". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.761374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf" (OuterVolumeSpecName: "server-conf") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.773424 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.803602 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" (UID: "75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830450 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830483 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhr7j\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-kube-api-access-vhr7j\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830526 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") on node \"crc\" " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830543 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830555 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830567 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830577 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830590 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.830601 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.848270 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.848432 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7c39547e-1168-404d-865b-13cfd1526f6f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f") on node "crc" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931356 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931431 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931497 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931632 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931665 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.931695 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") pod \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\" (UID: \"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d\") " Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932034 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932417 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932444 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.932459 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.935376 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info" (OuterVolumeSpecName: "pod-info") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.937568 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.937639 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v" (OuterVolumeSpecName: "kube-api-access-vsj5v") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "kube-api-access-vsj5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.942435 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2" (OuterVolumeSpecName: "persistence") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 15:27:19 crc kubenswrapper[4907]: I0313 15:27:19.957708 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf" (OuterVolumeSpecName: "server-conf") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.014692 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" (UID: "7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032393 4907 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032432 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032464 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") on node \"crc\" " Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032476 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032486 4907 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032495 4907 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032503 4907 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032515 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsj5v\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-kube-api-access-vsj5v\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.032523 4907 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.047122 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.047261 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2") on node "crc" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.133327 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.503061 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.574296 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.575745 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" containerID="cri-o://7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" gracePeriod=10 Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621299 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621320 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d","Type":"ContainerDied","Data":"7fa7486d25e0b1f49302f99040c36181735134e64166f989efe3f3ae26956d78"} Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.621415 4907 scope.go:117] "RemoveContainer" containerID="c2763a983f63aa1fe0a38c4e94c194af6a54549cd715d4e926bfe5396d362f36" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.672376 4907 scope.go:117] "RemoveContainer" containerID="bf96e8e63801badfacccb754ff56342cdc5ba8417a4f20f022b8a5f76d1c9272" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.682599 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.717770 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.725349 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.730063 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734623 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734907 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734923 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734936 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734942 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734961 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734969 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: E0313 15:27:20.734978 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.734984 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="setup-container" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.735102 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.735116 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" containerName="rabbitmq" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.741071 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.741176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.744867 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.745067 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.745204 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.745325 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-grjtp" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.752591 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.759084 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.765681 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.771230 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.771384 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.771490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-vlpfs" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.784160 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.785172 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.792699 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843683 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843701 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843738 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843762 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843821 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p486x\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-kube-api-access-p486x\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.843858 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945699 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945833 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzkdl\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-kube-api-access-mzkdl\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945898 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945934 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p486x\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-kube-api-access-p486x\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.945988 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946072 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946130 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946181 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946236 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946281 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.946301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.948268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.948717 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.948942 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.949036 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.952337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.954649 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.955743 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.955784 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d84a704eae1ca42fe329ab2f9d170d2a410be00de88c5e0a709d7b62965c26cb/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.956661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.965619 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p486x\" (UniqueName: \"kubernetes.io/projected/fc486d5d-ec94-461c-acf2-dcca08b2a1b6-kube-api-access-p486x\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:20 crc kubenswrapper[4907]: I0313 15:27:20.998306 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7c39547e-1168-404d-865b-13cfd1526f6f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7c39547e-1168-404d-865b-13cfd1526f6f\") pod \"rabbitmq-cell1-server-0\" (UID: \"fc486d5d-ec94-461c-acf2-dcca08b2a1b6\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.026235 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048373 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048487 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzkdl\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-kube-api-access-mzkdl\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048527 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048546 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.048995 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.049712 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.050182 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.051248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.054099 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.054139 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c926015d5702d92351724596ba892321797e624e5f7a01b00404b7d1c8a17788/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.056852 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.057748 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.064732 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.068211 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzkdl\" (UniqueName: \"kubernetes.io/projected/2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad-kube-api-access-mzkdl\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.075415 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.089316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9187efe6-871d-4bbe-9376-bff5d8a151f2\") pod \"rabbitmq-server-0\" (UID: \"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad\") " pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.099456 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.152482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") pod \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.152531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") pod \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.152628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") pod \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\" (UID: \"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2\") " Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.158809 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s" (OuterVolumeSpecName: "kube-api-access-r9x6s") pod "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" (UID: "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2"). InnerVolumeSpecName "kube-api-access-r9x6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.187622 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config" (OuterVolumeSpecName: "config") pod "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" (UID: "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.194957 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" (UID: "d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.254124 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9x6s\" (UniqueName: \"kubernetes.io/projected/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-kube-api-access-r9x6s\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.254165 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.254189 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.581070 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.587863 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 15:27:21 crc kubenswrapper[4907]: W0313 15:27:21.595853 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ae0c4fc_0b3a_47a3_a6db_3e0ae3d64dad.slice/crio-b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c WatchSource:0}: Error finding container b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c: Status 404 returned error can't find the container with id b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.635341 4907 generic.go:334] "Generic (PLEG): container finished" podID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" exitCode=0 Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636140 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636162 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerDied","Data":"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636735 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c95686bd5-8kb7m" event={"ID":"d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2","Type":"ContainerDied","Data":"18cfff04ebd6c2ab270b3fa1a352afaef9dc310f2e13e5df39289e1f0f968058"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.636761 4907 scope.go:117] "RemoveContainer" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.639444 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerStarted","Data":"694eb1066ead945b234eab0fd1a5c138170fd26599f8651b6d0198397d809e5a"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.641246 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerStarted","Data":"b99ed7eeb540a88d6a8940684577ebb52b99ec7a8bb229f5d1502b3a7666409c"} Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.673744 4907 scope.go:117] "RemoveContainer" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.722398 4907 scope.go:117] "RemoveContainer" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" Mar 13 15:27:21 crc kubenswrapper[4907]: E0313 15:27:21.723079 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea\": container with ID starting with 7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea not found: ID does not exist" containerID="7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.723219 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea"} err="failed to get container status \"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea\": rpc error: code = NotFound desc = could not find container \"7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea\": container with ID starting with 7dc8560cb6dc85650a7b469d68aaea22e66e3e209dca62e699bcb1b0807b2fea not found: ID does not exist" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.723351 4907 scope.go:117] "RemoveContainer" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" Mar 13 15:27:21 crc kubenswrapper[4907]: E0313 15:27:21.727374 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3\": container with ID starting with 8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3 not found: ID does not exist" containerID="8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.727450 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3"} err="failed to get container status \"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3\": rpc error: code = NotFound desc = could not find container \"8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3\": container with ID starting with 8f7fad973a21b1188eda2310b2c3e66fb866528b33ce2f78d5f9e1986ac222d3 not found: ID does not exist" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.756824 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.762869 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c95686bd5-8kb7m"] Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.793064 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd" path="/var/lib/kubelet/pods/75f775f2-b77c-4ff4-80e4-12f6d3e2d0fd/volumes" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.794811 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d" path="/var/lib/kubelet/pods/7cb75c96-2c96-4b6e-9390-3ca2f7e9fb0d/volumes" Mar 13 15:27:21 crc kubenswrapper[4907]: I0313 15:27:21.796267 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" path="/var/lib/kubelet/pods/d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2/volumes" Mar 13 15:27:23 crc kubenswrapper[4907]: I0313 15:27:23.662612 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerStarted","Data":"4e6bfc55b73ea14dd81f54b09e66186a0fe8a4af94eb0f240b0b9d243ebb9b45"} Mar 13 15:27:23 crc kubenswrapper[4907]: I0313 15:27:23.666052 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerStarted","Data":"9dc9569aced9e6a9d87eacfd4113edb23371b96cee87c58cb825f8e9ca84fd40"} Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.346155 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:37 crc kubenswrapper[4907]: E0313 15:27:37.346935 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="init" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.346946 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="init" Mar 13 15:27:37 crc kubenswrapper[4907]: E0313 15:27:37.346967 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.346973 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.347114 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b7f262-ac9a-4a7e-ae94-6ce13d4105a2" containerName="dnsmasq-dns" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.348106 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.361198 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.399200 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.399347 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.399368 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500391 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500807 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.500895 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.501083 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.523733 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"certified-operators-st569\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:37 crc kubenswrapper[4907]: I0313 15:27:37.666101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.122834 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.823306 4907 generic.go:334] "Generic (PLEG): container finished" podID="340af619-4557-47cd-89b6-86237b062783" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" exitCode=0 Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.823397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a"} Mar 13 15:27:38 crc kubenswrapper[4907]: I0313 15:27:38.823633 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerStarted","Data":"3e29c70762aed7c8ac6ec015bcd46c38075af57c46837c2bb98102a5838f0966"} Mar 13 15:27:39 crc kubenswrapper[4907]: I0313 15:27:39.834964 4907 generic.go:334] "Generic (PLEG): container finished" podID="340af619-4557-47cd-89b6-86237b062783" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" exitCode=0 Mar 13 15:27:39 crc kubenswrapper[4907]: I0313 15:27:39.835064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b"} Mar 13 15:27:40 crc kubenswrapper[4907]: I0313 15:27:40.848219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerStarted","Data":"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46"} Mar 13 15:27:40 crc kubenswrapper[4907]: I0313 15:27:40.874534 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-st569" podStartSLOduration=2.46863519 podStartE2EDuration="3.874505066s" podCreationTimestamp="2026-03-13 15:27:37 +0000 UTC" firstStartedPulling="2026-03-13 15:27:38.825861787 +0000 UTC m=+4957.725649476" lastFinishedPulling="2026-03-13 15:27:40.231731663 +0000 UTC m=+4959.131519352" observedRunningTime="2026-03-13 15:27:40.86731209 +0000 UTC m=+4959.767099819" watchObservedRunningTime="2026-03-13 15:27:40.874505066 +0000 UTC m=+4959.774292795" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.666436 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.667282 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.704465 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.943410 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:47 crc kubenswrapper[4907]: I0313 15:27:47.996603 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:49 crc kubenswrapper[4907]: I0313 15:27:49.918195 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-st569" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" containerID="cri-o://77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" gracePeriod=2 Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.280235 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.398137 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") pod \"340af619-4557-47cd-89b6-86237b062783\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.398247 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") pod \"340af619-4557-47cd-89b6-86237b062783\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.398281 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") pod \"340af619-4557-47cd-89b6-86237b062783\" (UID: \"340af619-4557-47cd-89b6-86237b062783\") " Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.400199 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities" (OuterVolumeSpecName: "utilities") pod "340af619-4557-47cd-89b6-86237b062783" (UID: "340af619-4557-47cd-89b6-86237b062783"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.403332 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf" (OuterVolumeSpecName: "kube-api-access-9pqvf") pod "340af619-4557-47cd-89b6-86237b062783" (UID: "340af619-4557-47cd-89b6-86237b062783"). InnerVolumeSpecName "kube-api-access-9pqvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.500998 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.501295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pqvf\" (UniqueName: \"kubernetes.io/projected/340af619-4557-47cd-89b6-86237b062783-kube-api-access-9pqvf\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932658 4907 generic.go:334] "Generic (PLEG): container finished" podID="340af619-4557-47cd-89b6-86237b062783" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" exitCode=0 Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46"} Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932799 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-st569" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932830 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-st569" event={"ID":"340af619-4557-47cd-89b6-86237b062783","Type":"ContainerDied","Data":"3e29c70762aed7c8ac6ec015bcd46c38075af57c46837c2bb98102a5838f0966"} Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.932952 4907 scope.go:117] "RemoveContainer" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" Mar 13 15:27:50 crc kubenswrapper[4907]: I0313 15:27:50.958685 4907 scope.go:117] "RemoveContainer" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.001416 4907 scope.go:117] "RemoveContainer" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.025770 4907 scope.go:117] "RemoveContainer" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" Mar 13 15:27:51 crc kubenswrapper[4907]: E0313 15:27:51.026277 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46\": container with ID starting with 77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46 not found: ID does not exist" containerID="77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.026323 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46"} err="failed to get container status \"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46\": rpc error: code = NotFound desc = could not find container \"77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46\": container with ID starting with 77a6cca1c6bd5dff35a26d10feca15c2f38dbcf318403240f5c7990157c4ab46 not found: ID does not exist" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.026357 4907 scope.go:117] "RemoveContainer" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" Mar 13 15:27:51 crc kubenswrapper[4907]: E0313 15:27:51.027017 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b\": container with ID starting with e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b not found: ID does not exist" containerID="e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.027073 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b"} err="failed to get container status \"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b\": rpc error: code = NotFound desc = could not find container \"e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b\": container with ID starting with e7e8d7b2fd73572e0983d90f865c19570898239292fc89f15261b7908adb8b0b not found: ID does not exist" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.027096 4907 scope.go:117] "RemoveContainer" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" Mar 13 15:27:51 crc kubenswrapper[4907]: E0313 15:27:51.027418 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a\": container with ID starting with 1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a not found: ID does not exist" containerID="1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.027454 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a"} err="failed to get container status \"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a\": rpc error: code = NotFound desc = could not find container \"1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a\": container with ID starting with 1e3d0562c691019dbf46762447ac1cd38a0e5e62d058f6ab0fe85ff6f764b47a not found: ID does not exist" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.304544 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "340af619-4557-47cd-89b6-86237b062783" (UID: "340af619-4557-47cd-89b6-86237b062783"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.313283 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/340af619-4557-47cd-89b6-86237b062783-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.566559 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.575756 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-st569"] Mar 13 15:27:51 crc kubenswrapper[4907]: I0313 15:27:51.795859 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="340af619-4557-47cd-89b6-86237b062783" path="/var/lib/kubelet/pods/340af619-4557-47cd-89b6-86237b062783/volumes" Mar 13 15:27:54 crc kubenswrapper[4907]: I0313 15:27:54.964574 4907 generic.go:334] "Generic (PLEG): container finished" podID="fc486d5d-ec94-461c-acf2-dcca08b2a1b6" containerID="9dc9569aced9e6a9d87eacfd4113edb23371b96cee87c58cb825f8e9ca84fd40" exitCode=0 Mar 13 15:27:54 crc kubenswrapper[4907]: I0313 15:27:54.964688 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerDied","Data":"9dc9569aced9e6a9d87eacfd4113edb23371b96cee87c58cb825f8e9ca84fd40"} Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.976213 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fc486d5d-ec94-461c-acf2-dcca08b2a1b6","Type":"ContainerStarted","Data":"49d332f53efdfc23759dba52b6372c5d1993e483aca345df376f98c207b6db3b"} Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.977043 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.980815 4907 generic.go:334] "Generic (PLEG): container finished" podID="2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad" containerID="4e6bfc55b73ea14dd81f54b09e66186a0fe8a4af94eb0f240b0b9d243ebb9b45" exitCode=0 Mar 13 15:27:55 crc kubenswrapper[4907]: I0313 15:27:55.980867 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerDied","Data":"4e6bfc55b73ea14dd81f54b09e66186a0fe8a4af94eb0f240b0b9d243ebb9b45"} Mar 13 15:27:56 crc kubenswrapper[4907]: I0313 15:27:56.007434 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.007405306 podStartE2EDuration="36.007405306s" podCreationTimestamp="2026-03-13 15:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:27:56.005860813 +0000 UTC m=+4974.905648502" watchObservedRunningTime="2026-03-13 15:27:56.007405306 +0000 UTC m=+4974.907193015" Mar 13 15:27:56 crc kubenswrapper[4907]: I0313 15:27:56.989622 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad","Type":"ContainerStarted","Data":"99e651808a9607da115a9154bac22f68298b64908ab45c28005107610dea2392"} Mar 13 15:27:56 crc kubenswrapper[4907]: I0313 15:27:56.990196 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 15:27:57 crc kubenswrapper[4907]: I0313 15:27:57.017231 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.017208514 podStartE2EDuration="37.017208514s" podCreationTimestamp="2026-03-13 15:27:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:27:57.013375969 +0000 UTC m=+4975.913163698" watchObservedRunningTime="2026-03-13 15:27:57.017208514 +0000 UTC m=+4975.916996213" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143154 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:28:00 crc kubenswrapper[4907]: E0313 15:28:00.143793 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-content" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143807 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-content" Mar 13 15:28:00 crc kubenswrapper[4907]: E0313 15:28:00.143816 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-utilities" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143823 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="extract-utilities" Mar 13 15:28:00 crc kubenswrapper[4907]: E0313 15:28:00.143837 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.143843 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.144007 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="340af619-4557-47cd-89b6-86237b062783" containerName="registry-server" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.144514 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.146841 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.146849 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.153117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.154759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.240835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"auto-csr-approver-29556928-nh6lw\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.342228 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"auto-csr-approver-29556928-nh6lw\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.361821 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"auto-csr-approver-29556928-nh6lw\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.461179 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:00 crc kubenswrapper[4907]: I0313 15:28:00.885099 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:28:00 crc kubenswrapper[4907]: W0313 15:28:00.892441 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod566ddaa7_ea07_421b_a181_0aa768454ba6.slice/crio-e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7 WatchSource:0}: Error finding container e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7: Status 404 returned error can't find the container with id e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7 Mar 13 15:28:01 crc kubenswrapper[4907]: I0313 15:28:01.014716 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" event={"ID":"566ddaa7-ea07-421b-a181-0aa768454ba6","Type":"ContainerStarted","Data":"e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7"} Mar 13 15:28:03 crc kubenswrapper[4907]: I0313 15:28:03.028840 4907 generic.go:334] "Generic (PLEG): container finished" podID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerID="206bf50c83686d14437a447b0be4276ceccf75f1f0402647af114c3e1ce67140" exitCode=0 Mar 13 15:28:03 crc kubenswrapper[4907]: I0313 15:28:03.029128 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" event={"ID":"566ddaa7-ea07-421b-a181-0aa768454ba6","Type":"ContainerDied","Data":"206bf50c83686d14437a447b0be4276ceccf75f1f0402647af114c3e1ce67140"} Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.291257 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.407480 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") pod \"566ddaa7-ea07-421b-a181-0aa768454ba6\" (UID: \"566ddaa7-ea07-421b-a181-0aa768454ba6\") " Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.412714 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr" (OuterVolumeSpecName: "kube-api-access-5l9fr") pod "566ddaa7-ea07-421b-a181-0aa768454ba6" (UID: "566ddaa7-ea07-421b-a181-0aa768454ba6"). InnerVolumeSpecName "kube-api-access-5l9fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:28:04 crc kubenswrapper[4907]: I0313 15:28:04.510077 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l9fr\" (UniqueName: \"kubernetes.io/projected/566ddaa7-ea07-421b-a181-0aa768454ba6-kube-api-access-5l9fr\") on node \"crc\" DevicePath \"\"" Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.045938 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" event={"ID":"566ddaa7-ea07-421b-a181-0aa768454ba6","Type":"ContainerDied","Data":"e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7"} Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.045978 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2be2e36196a0df0aa7aaeb1f881a8eebe3037ab79d154d7fe5a1e31ed4297b7" Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.046007 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556928-nh6lw" Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.362002 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.366756 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556922-b7qtk"] Mar 13 15:28:05 crc kubenswrapper[4907]: I0313 15:28:05.791941 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31cb9010-2aab-475d-aab3-9aa35a538581" path="/var/lib/kubelet/pods/31cb9010-2aab-475d-aab3-9aa35a538581/volumes" Mar 13 15:28:11 crc kubenswrapper[4907]: I0313 15:28:11.078132 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 15:28:11 crc kubenswrapper[4907]: I0313 15:28:11.102155 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.942598 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:22 crc kubenswrapper[4907]: E0313 15:28:22.943554 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerName="oc" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.943573 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerName="oc" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.943732 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" containerName="oc" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.944409 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.946781 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mqs4s" Mar 13 15:28:22 crc kubenswrapper[4907]: I0313 15:28:22.952315 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.087819 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"mariadb-client\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.189588 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"mariadb-client\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.207396 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"mariadb-client\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.269528 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:23 crc kubenswrapper[4907]: I0313 15:28:23.766961 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:24 crc kubenswrapper[4907]: I0313 15:28:24.195494 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerStarted","Data":"82d4fd4ae31f9ad44ead28b840a4b0ae479a0cec913d3bdb2dc24a7f47cb13e9"} Mar 13 15:28:25 crc kubenswrapper[4907]: I0313 15:28:25.207828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerStarted","Data":"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37"} Mar 13 15:28:25 crc kubenswrapper[4907]: I0313 15:28:25.244349 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client" podStartSLOduration=2.708066529 podStartE2EDuration="3.244327757s" podCreationTimestamp="2026-03-13 15:28:22 +0000 UTC" firstStartedPulling="2026-03-13 15:28:23.771782792 +0000 UTC m=+5002.671570481" lastFinishedPulling="2026-03-13 15:28:24.30804402 +0000 UTC m=+5003.207831709" observedRunningTime="2026-03-13 15:28:25.236061001 +0000 UTC m=+5004.135848760" watchObservedRunningTime="2026-03-13 15:28:25.244327757 +0000 UTC m=+5004.144115456" Mar 13 15:28:27 crc kubenswrapper[4907]: I0313 15:28:27.418291 4907 scope.go:117] "RemoveContainer" containerID="2cd2416cb09c17cb001811639f27834ed34890771029fd10f4c0a96794fa7519" Mar 13 15:28:27 crc kubenswrapper[4907]: I0313 15:28:27.480644 4907 scope.go:117] "RemoveContainer" containerID="d45e59416766d20c4f95b995286178dd988b4b6257768b1fef8c82dcc281cd6c" Mar 13 15:28:38 crc kubenswrapper[4907]: I0313 15:28:38.507483 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:38 crc kubenswrapper[4907]: I0313 15:28:38.508436 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-client" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" containerID="cri-o://076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" gracePeriod=30 Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.030527 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.145935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") pod \"c39974e7-120a-410b-b78f-d3588b2cc22d\" (UID: \"c39974e7-120a-410b-b78f-d3588b2cc22d\") " Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.153152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7" (OuterVolumeSpecName: "kube-api-access-sc7c7") pod "c39974e7-120a-410b-b78f-d3588b2cc22d" (UID: "c39974e7-120a-410b-b78f-d3588b2cc22d"). InnerVolumeSpecName "kube-api-access-sc7c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.247753 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc7c7\" (UniqueName: \"kubernetes.io/projected/c39974e7-120a-410b-b78f-d3588b2cc22d-kube-api-access-sc7c7\") on node \"crc\" DevicePath \"\"" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320913 4907 generic.go:334] "Generic (PLEG): container finished" podID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" exitCode=143 Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320953 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320965 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerDied","Data":"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37"} Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.320988 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c39974e7-120a-410b-b78f-d3588b2cc22d","Type":"ContainerDied","Data":"82d4fd4ae31f9ad44ead28b840a4b0ae479a0cec913d3bdb2dc24a7f47cb13e9"} Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.321004 4907 scope.go:117] "RemoveContainer" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.338840 4907 scope.go:117] "RemoveContainer" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" Mar 13 15:28:39 crc kubenswrapper[4907]: E0313 15:28:39.339212 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37\": container with ID starting with 076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37 not found: ID does not exist" containerID="076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.339243 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37"} err="failed to get container status \"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37\": rpc error: code = NotFound desc = could not find container \"076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37\": container with ID starting with 076daa9876f3ca96d1b1fb0a5e45b1cf807c92541a65b07004b19037cc101e37 not found: ID does not exist" Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.353282 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.356180 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:28:39 crc kubenswrapper[4907]: I0313 15:28:39.797426 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" path="/var/lib/kubelet/pods/c39974e7-120a-410b-b78f-d3588b2cc22d/volumes" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.762440 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:28:40 crc kubenswrapper[4907]: E0313 15:28:40.762815 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.762831 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.763031 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39974e7-120a-410b-b78f-d3588b2cc22d" containerName="mariadb-client" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.764286 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.775052 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.868935 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.868993 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.869066 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970568 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970615 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.970979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:40 crc kubenswrapper[4907]: I0313 15:28:40.971052 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:41 crc kubenswrapper[4907]: I0313 15:28:41.003278 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"redhat-operators-jzxtj\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:41 crc kubenswrapper[4907]: I0313 15:28:41.082981 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:41 crc kubenswrapper[4907]: I0313 15:28:41.518568 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:28:42 crc kubenswrapper[4907]: I0313 15:28:42.347133 4907 generic.go:334] "Generic (PLEG): container finished" podID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" exitCode=0 Mar 13 15:28:42 crc kubenswrapper[4907]: I0313 15:28:42.347185 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239"} Mar 13 15:28:42 crc kubenswrapper[4907]: I0313 15:28:42.347212 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerStarted","Data":"402cc4345879db4aec2b048632e5651fc97bb6231de9ed25658ecbdd0404e855"} Mar 13 15:28:44 crc kubenswrapper[4907]: I0313 15:28:44.365126 4907 generic.go:334] "Generic (PLEG): container finished" podID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" exitCode=0 Mar 13 15:28:44 crc kubenswrapper[4907]: I0313 15:28:44.365174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7"} Mar 13 15:28:45 crc kubenswrapper[4907]: I0313 15:28:45.375916 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerStarted","Data":"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312"} Mar 13 15:28:45 crc kubenswrapper[4907]: I0313 15:28:45.402486 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jzxtj" podStartSLOduration=2.96754561 podStartE2EDuration="5.402465886s" podCreationTimestamp="2026-03-13 15:28:40 +0000 UTC" firstStartedPulling="2026-03-13 15:28:42.348805534 +0000 UTC m=+5021.248593223" lastFinishedPulling="2026-03-13 15:28:44.78372581 +0000 UTC m=+5023.683513499" observedRunningTime="2026-03-13 15:28:45.395032903 +0000 UTC m=+5024.294820592" watchObservedRunningTime="2026-03-13 15:28:45.402465886 +0000 UTC m=+5024.302253575" Mar 13 15:28:48 crc kubenswrapper[4907]: I0313 15:28:48.042245 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:28:48 crc kubenswrapper[4907]: I0313 15:28:48.042821 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:28:51 crc kubenswrapper[4907]: I0313 15:28:51.083955 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:51 crc kubenswrapper[4907]: I0313 15:28:51.085154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:28:52 crc kubenswrapper[4907]: I0313 15:28:52.126839 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jzxtj" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" probeResult="failure" output=< Mar 13 15:28:52 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:28:52 crc kubenswrapper[4907]: > Mar 13 15:29:01 crc kubenswrapper[4907]: I0313 15:29:01.130849 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:01 crc kubenswrapper[4907]: I0313 15:29:01.275580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:01 crc kubenswrapper[4907]: I0313 15:29:01.383328 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:29:02 crc kubenswrapper[4907]: I0313 15:29:02.492482 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jzxtj" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" containerID="cri-o://23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" gracePeriod=2 Mar 13 15:29:02 crc kubenswrapper[4907]: I0313 15:29:02.895638 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.005631 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") pod \"471a271b-2d62-4bdb-956d-91e268d2e4b5\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.005717 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") pod \"471a271b-2d62-4bdb-956d-91e268d2e4b5\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.005846 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") pod \"471a271b-2d62-4bdb-956d-91e268d2e4b5\" (UID: \"471a271b-2d62-4bdb-956d-91e268d2e4b5\") " Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.006503 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities" (OuterVolumeSpecName: "utilities") pod "471a271b-2d62-4bdb-956d-91e268d2e4b5" (UID: "471a271b-2d62-4bdb-956d-91e268d2e4b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.010741 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj" (OuterVolumeSpecName: "kube-api-access-g76wj") pod "471a271b-2d62-4bdb-956d-91e268d2e4b5" (UID: "471a271b-2d62-4bdb-956d-91e268d2e4b5"). InnerVolumeSpecName "kube-api-access-g76wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.107640 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.107704 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g76wj\" (UniqueName: \"kubernetes.io/projected/471a271b-2d62-4bdb-956d-91e268d2e4b5-kube-api-access-g76wj\") on node \"crc\" DevicePath \"\"" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.133723 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "471a271b-2d62-4bdb-956d-91e268d2e4b5" (UID: "471a271b-2d62-4bdb-956d-91e268d2e4b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.209141 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/471a271b-2d62-4bdb-956d-91e268d2e4b5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503124 4907 generic.go:334] "Generic (PLEG): container finished" podID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" exitCode=0 Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312"} Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503202 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jzxtj" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503224 4907 scope.go:117] "RemoveContainer" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.503213 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jzxtj" event={"ID":"471a271b-2d62-4bdb-956d-91e268d2e4b5","Type":"ContainerDied","Data":"402cc4345879db4aec2b048632e5651fc97bb6231de9ed25658ecbdd0404e855"} Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.520522 4907 scope.go:117] "RemoveContainer" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.542562 4907 scope.go:117] "RemoveContainer" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.577560 4907 scope.go:117] "RemoveContainer" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" Mar 13 15:29:03 crc kubenswrapper[4907]: E0313 15:29:03.578261 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312\": container with ID starting with 23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312 not found: ID does not exist" containerID="23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578320 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312"} err="failed to get container status \"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312\": rpc error: code = NotFound desc = could not find container \"23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312\": container with ID starting with 23e6b0d481bf70d0a2451207b0ebb74ac5446201d8e102faf3f204db678a0312 not found: ID does not exist" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578358 4907 scope.go:117] "RemoveContainer" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" Mar 13 15:29:03 crc kubenswrapper[4907]: E0313 15:29:03.578813 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7\": container with ID starting with 6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7 not found: ID does not exist" containerID="6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578836 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7"} err="failed to get container status \"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7\": rpc error: code = NotFound desc = could not find container \"6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7\": container with ID starting with 6aab4dd9c3bca6c8c51dfdb1e941984b5dcd418ed1ef52f4155b9cfe47b28ac7 not found: ID does not exist" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.578852 4907 scope.go:117] "RemoveContainer" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" Mar 13 15:29:03 crc kubenswrapper[4907]: E0313 15:29:03.579501 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239\": container with ID starting with 717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239 not found: ID does not exist" containerID="717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.579540 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239"} err="failed to get container status \"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239\": rpc error: code = NotFound desc = could not find container \"717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239\": container with ID starting with 717b6dbc97d7ec5e02f0d19fcad760520fac7eb81149b49eb75a2be915500239 not found: ID does not exist" Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.580162 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.586025 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jzxtj"] Mar 13 15:29:03 crc kubenswrapper[4907]: I0313 15:29:03.802039 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" path="/var/lib/kubelet/pods/471a271b-2d62-4bdb-956d-91e268d2e4b5/volumes" Mar 13 15:29:18 crc kubenswrapper[4907]: I0313 15:29:18.041368 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:29:18 crc kubenswrapper[4907]: I0313 15:29:18.042139 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.041575 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042139 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042181 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042642 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.042684 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" gracePeriod=600 Mar 13 15:29:48 crc kubenswrapper[4907]: E0313 15:29:48.165022 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.828922 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" exitCode=0 Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.828970 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596"} Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.829008 4907 scope.go:117] "RemoveContainer" containerID="b5bb3cf2979988d507615c0a856eebc9a02c4d2c3275ceb2cee45dcc6b625cfd" Mar 13 15:29:48 crc kubenswrapper[4907]: I0313 15:29:48.829659 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:29:48 crc kubenswrapper[4907]: E0313 15:29:48.830036 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.145461 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.146402 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-utilities" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146420 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-utilities" Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.146439 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146447 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.146475 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-content" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146483 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="extract-content" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.146648 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="471a271b-2d62-4bdb-956d-91e268d2e4b5" containerName="registry-server" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.147330 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.149802 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.151367 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.156527 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.162371 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.176236 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.177772 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.181739 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.193099 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.197838 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269147 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269203 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"auto-csr-approver-29556930-ndtqs\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269235 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.269537 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"auto-csr-approver-29556930-ndtqs\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.371227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.372155 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.378200 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.390217 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"auto-csr-approver-29556930-ndtqs\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.390304 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"collect-profiles-29556930-f852z\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.470631 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.490827 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.782726 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:00 crc kubenswrapper[4907]: E0313 15:30:00.783154 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.905108 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.929026 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" event={"ID":"2983956e-5d8d-4595-9fda-91759741010a","Type":"ContainerStarted","Data":"0617586958aed9d4fa53068330454c7a10480fc0c59b1ca5cc83cc8bc439a18b"} Mar 13 15:30:00 crc kubenswrapper[4907]: I0313 15:30:00.964766 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:30:00 crc kubenswrapper[4907]: W0313 15:30:00.967138 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf83f491_eb7d_4e1c_b814_28cc7ef9abee.slice/crio-3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591 WatchSource:0}: Error finding container 3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591: Status 404 returned error can't find the container with id 3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591 Mar 13 15:30:01 crc kubenswrapper[4907]: I0313 15:30:01.936454 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" event={"ID":"df83f491-eb7d-4e1c-b814-28cc7ef9abee","Type":"ContainerStarted","Data":"3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591"} Mar 13 15:30:01 crc kubenswrapper[4907]: I0313 15:30:01.937728 4907 generic.go:334] "Generic (PLEG): container finished" podID="2983956e-5d8d-4595-9fda-91759741010a" containerID="74d889f05e8d8fa3b0cfc7fed167cff28426e276bf0f0c9f759dd1496f3c1058" exitCode=0 Mar 13 15:30:01 crc kubenswrapper[4907]: I0313 15:30:01.937773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" event={"ID":"2983956e-5d8d-4595-9fda-91759741010a","Type":"ContainerDied","Data":"74d889f05e8d8fa3b0cfc7fed167cff28426e276bf0f0c9f759dd1496f3c1058"} Mar 13 15:30:02 crc kubenswrapper[4907]: I0313 15:30:02.946778 4907 generic.go:334] "Generic (PLEG): container finished" podID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerID="9c966e9f83d2d89ea300010e1e5e03c7e34fbdc9099449a4ffcb481299d87028" exitCode=0 Mar 13 15:30:02 crc kubenswrapper[4907]: I0313 15:30:02.947416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" event={"ID":"df83f491-eb7d-4e1c-b814-28cc7ef9abee","Type":"ContainerDied","Data":"9c966e9f83d2d89ea300010e1e5e03c7e34fbdc9099449a4ffcb481299d87028"} Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.219736 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.314114 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") pod \"2983956e-5d8d-4595-9fda-91759741010a\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.314162 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") pod \"2983956e-5d8d-4595-9fda-91759741010a\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.314205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") pod \"2983956e-5d8d-4595-9fda-91759741010a\" (UID: \"2983956e-5d8d-4595-9fda-91759741010a\") " Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.315190 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume" (OuterVolumeSpecName: "config-volume") pod "2983956e-5d8d-4595-9fda-91759741010a" (UID: "2983956e-5d8d-4595-9fda-91759741010a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.320696 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2983956e-5d8d-4595-9fda-91759741010a" (UID: "2983956e-5d8d-4595-9fda-91759741010a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.321215 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh" (OuterVolumeSpecName: "kube-api-access-jrrsh") pod "2983956e-5d8d-4595-9fda-91759741010a" (UID: "2983956e-5d8d-4595-9fda-91759741010a"). InnerVolumeSpecName "kube-api-access-jrrsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.416183 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrrsh\" (UniqueName: \"kubernetes.io/projected/2983956e-5d8d-4595-9fda-91759741010a-kube-api-access-jrrsh\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.416230 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2983956e-5d8d-4595-9fda-91759741010a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.416244 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2983956e-5d8d-4595-9fda-91759741010a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.955929 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.955930 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z" event={"ID":"2983956e-5d8d-4595-9fda-91759741010a","Type":"ContainerDied","Data":"0617586958aed9d4fa53068330454c7a10480fc0c59b1ca5cc83cc8bc439a18b"} Mar 13 15:30:03 crc kubenswrapper[4907]: I0313 15:30:03.956353 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0617586958aed9d4fa53068330454c7a10480fc0c59b1ca5cc83cc8bc439a18b" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.233925 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.303910 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.319139 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556885-9xsw6"] Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.328832 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") pod \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\" (UID: \"df83f491-eb7d-4e1c-b814-28cc7ef9abee\") " Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.334088 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p" (OuterVolumeSpecName: "kube-api-access-ctk6p") pod "df83f491-eb7d-4e1c-b814-28cc7ef9abee" (UID: "df83f491-eb7d-4e1c-b814-28cc7ef9abee"). InnerVolumeSpecName "kube-api-access-ctk6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.430550 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctk6p\" (UniqueName: \"kubernetes.io/projected/df83f491-eb7d-4e1c-b814-28cc7ef9abee-kube-api-access-ctk6p\") on node \"crc\" DevicePath \"\"" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.964735 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" event={"ID":"df83f491-eb7d-4e1c-b814-28cc7ef9abee","Type":"ContainerDied","Data":"3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591"} Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.965041 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3091fd9a1e1aec73474d08add55adf04d2c04dd62f815e3f3b5c66e65aca9591" Mar 13 15:30:04 crc kubenswrapper[4907]: I0313 15:30:04.964776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556930-ndtqs" Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.295617 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.303218 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556924-8z5rs"] Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.793529 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a50278-3da1-4a63-ac7e-1fcb4ae77f52" path="/var/lib/kubelet/pods/11a50278-3da1-4a63-ac7e-1fcb4ae77f52/volumes" Mar 13 15:30:05 crc kubenswrapper[4907]: I0313 15:30:05.794645 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d79a5eae-26dc-4f69-aa63-9bd6c5788a03" path="/var/lib/kubelet/pods/d79a5eae-26dc-4f69-aa63-9bd6c5788a03/volumes" Mar 13 15:30:11 crc kubenswrapper[4907]: I0313 15:30:11.786140 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:11 crc kubenswrapper[4907]: E0313 15:30:11.787658 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:24 crc kubenswrapper[4907]: I0313 15:30:24.783160 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:24 crc kubenswrapper[4907]: E0313 15:30:24.783825 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:27 crc kubenswrapper[4907]: I0313 15:30:27.604319 4907 scope.go:117] "RemoveContainer" containerID="34346eaf9fc976e698b78600d581550e67e1b840574df19f7129d1f617293a7d" Mar 13 15:30:27 crc kubenswrapper[4907]: I0313 15:30:27.633308 4907 scope.go:117] "RemoveContainer" containerID="16a638d449eaf9553071c678dcf92f038cd2ddde8f659099ade22f5e78a0fcae" Mar 13 15:30:38 crc kubenswrapper[4907]: I0313 15:30:38.782687 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:38 crc kubenswrapper[4907]: E0313 15:30:38.783366 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:30:50 crc kubenswrapper[4907]: I0313 15:30:50.782797 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:30:50 crc kubenswrapper[4907]: E0313 15:30:50.783618 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:01 crc kubenswrapper[4907]: I0313 15:31:01.782373 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:01 crc kubenswrapper[4907]: E0313 15:31:01.783033 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:13 crc kubenswrapper[4907]: I0313 15:31:13.782674 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:13 crc kubenswrapper[4907]: E0313 15:31:13.783426 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:27 crc kubenswrapper[4907]: I0313 15:31:27.782572 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:27 crc kubenswrapper[4907]: E0313 15:31:27.785538 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:42 crc kubenswrapper[4907]: I0313 15:31:42.782927 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:42 crc kubenswrapper[4907]: E0313 15:31:42.783725 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:31:54 crc kubenswrapper[4907]: I0313 15:31:54.783389 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:31:54 crc kubenswrapper[4907]: E0313 15:31:54.784226 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.136504 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:32:00 crc kubenswrapper[4907]: E0313 15:32:00.138788 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerName="oc" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.138970 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerName="oc" Mar 13 15:32:00 crc kubenswrapper[4907]: E0313 15:32:00.139100 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2983956e-5d8d-4595-9fda-91759741010a" containerName="collect-profiles" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.139183 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2983956e-5d8d-4595-9fda-91759741010a" containerName="collect-profiles" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.139433 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" containerName="oc" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.139534 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2983956e-5d8d-4595-9fda-91759741010a" containerName="collect-profiles" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.140302 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.142917 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.143102 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.143566 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.145282 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.324159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"auto-csr-approver-29556932-fxhnm\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.426076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"auto-csr-approver-29556932-fxhnm\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.445999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"auto-csr-approver-29556932-fxhnm\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.474024 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.897674 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:32:00 crc kubenswrapper[4907]: I0313 15:32:00.905683 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:32:01 crc kubenswrapper[4907]: I0313 15:32:01.090957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerStarted","Data":"8ff7e377b3f47cd070a55919264b03011577973ad91350e5e107b8bfd0e083f2"} Mar 13 15:32:02 crc kubenswrapper[4907]: I0313 15:32:02.101733 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerStarted","Data":"44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e"} Mar 13 15:32:02 crc kubenswrapper[4907]: I0313 15:32:02.120167 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" podStartSLOduration=1.206299333 podStartE2EDuration="2.120142975s" podCreationTimestamp="2026-03-13 15:32:00 +0000 UTC" firstStartedPulling="2026-03-13 15:32:00.905443528 +0000 UTC m=+5219.805231217" lastFinishedPulling="2026-03-13 15:32:01.81928714 +0000 UTC m=+5220.719074859" observedRunningTime="2026-03-13 15:32:02.11190537 +0000 UTC m=+5221.011693089" watchObservedRunningTime="2026-03-13 15:32:02.120142975 +0000 UTC m=+5221.019930664" Mar 13 15:32:03 crc kubenswrapper[4907]: I0313 15:32:03.111078 4907 generic.go:334] "Generic (PLEG): container finished" podID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerID="44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e" exitCode=0 Mar 13 15:32:03 crc kubenswrapper[4907]: I0313 15:32:03.111236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerDied","Data":"44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e"} Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.405749 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.491700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") pod \"e5efd747-3795-4922-9f81-3ef02dcb2f91\" (UID: \"e5efd747-3795-4922-9f81-3ef02dcb2f91\") " Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.496971 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8" (OuterVolumeSpecName: "kube-api-access-n2cs8") pod "e5efd747-3795-4922-9f81-3ef02dcb2f91" (UID: "e5efd747-3795-4922-9f81-3ef02dcb2f91"). InnerVolumeSpecName "kube-api-access-n2cs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.593427 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2cs8\" (UniqueName: \"kubernetes.io/projected/e5efd747-3795-4922-9f81-3ef02dcb2f91-kube-api-access-n2cs8\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.886954 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:32:04 crc kubenswrapper[4907]: I0313 15:32:04.896433 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556926-r9tzc"] Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.131074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" event={"ID":"e5efd747-3795-4922-9f81-3ef02dcb2f91","Type":"ContainerDied","Data":"8ff7e377b3f47cd070a55919264b03011577973ad91350e5e107b8bfd0e083f2"} Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.131148 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556932-fxhnm" Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.131157 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ff7e377b3f47cd070a55919264b03011577973ad91350e5e107b8bfd0e083f2" Mar 13 15:32:05 crc kubenswrapper[4907]: I0313 15:32:05.791565 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81a286dc-1af0-43f3-ad06-0e7d42141a0a" path="/var/lib/kubelet/pods/81a286dc-1af0-43f3-ad06-0e7d42141a0a/volumes" Mar 13 15:32:06 crc kubenswrapper[4907]: I0313 15:32:06.783279 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:06 crc kubenswrapper[4907]: E0313 15:32:06.783570 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:21 crc kubenswrapper[4907]: I0313 15:32:21.787439 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:21 crc kubenswrapper[4907]: E0313 15:32:21.789211 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:27 crc kubenswrapper[4907]: I0313 15:32:27.744104 4907 scope.go:117] "RemoveContainer" containerID="0b4af08a62b26495d3442675a1ce03aa549cddc0c7a28e8f6cb98a7deb792d53" Mar 13 15:32:27 crc kubenswrapper[4907]: I0313 15:32:27.765128 4907 scope.go:117] "RemoveContainer" containerID="fdb72e54eff786808e8d908105afbd91c2b9b22c84cb13b6b15ad7f081283322" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.301579 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:31 crc kubenswrapper[4907]: E0313 15:32:31.302306 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerName="oc" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.302320 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerName="oc" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.302472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" containerName="oc" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.303634 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.316550 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.441455 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.441599 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.441644 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.543625 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.543690 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.543747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.544406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.544461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.565693 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"redhat-marketplace-5nvvj\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.621444 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:31 crc kubenswrapper[4907]: I0313 15:32:31.882285 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:32 crc kubenswrapper[4907]: I0313 15:32:32.318153 4907 generic.go:334] "Generic (PLEG): container finished" podID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" exitCode=0 Mar 13 15:32:32 crc kubenswrapper[4907]: I0313 15:32:32.318237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b"} Mar 13 15:32:32 crc kubenswrapper[4907]: I0313 15:32:32.318506 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerStarted","Data":"b9169ee22d40d10babc6f49b1bfe9a71b6b37f17774a82002ce20e9b46ed79d4"} Mar 13 15:32:33 crc kubenswrapper[4907]: I0313 15:32:33.328297 4907 generic.go:334] "Generic (PLEG): container finished" podID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" exitCode=0 Mar 13 15:32:33 crc kubenswrapper[4907]: I0313 15:32:33.328396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50"} Mar 13 15:32:34 crc kubenswrapper[4907]: I0313 15:32:34.341582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerStarted","Data":"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182"} Mar 13 15:32:34 crc kubenswrapper[4907]: I0313 15:32:34.360699 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5nvvj" podStartSLOduration=1.937383814 podStartE2EDuration="3.360677472s" podCreationTimestamp="2026-03-13 15:32:31 +0000 UTC" firstStartedPulling="2026-03-13 15:32:32.320031367 +0000 UTC m=+5251.219819046" lastFinishedPulling="2026-03-13 15:32:33.743325015 +0000 UTC m=+5252.643112704" observedRunningTime="2026-03-13 15:32:34.357287449 +0000 UTC m=+5253.257075148" watchObservedRunningTime="2026-03-13 15:32:34.360677472 +0000 UTC m=+5253.260465171" Mar 13 15:32:35 crc kubenswrapper[4907]: I0313 15:32:35.782476 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:35 crc kubenswrapper[4907]: E0313 15:32:35.783106 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:41 crc kubenswrapper[4907]: I0313 15:32:41.621750 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:41 crc kubenswrapper[4907]: I0313 15:32:41.622343 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:41 crc kubenswrapper[4907]: I0313 15:32:41.668490 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:42 crc kubenswrapper[4907]: I0313 15:32:42.441739 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:42 crc kubenswrapper[4907]: I0313 15:32:42.493135 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.415796 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5nvvj" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" containerID="cri-o://28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" gracePeriod=2 Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.850518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.939780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") pod \"94531b92-db3d-4e91-acbc-ce321d2363a6\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.939841 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") pod \"94531b92-db3d-4e91-acbc-ce321d2363a6\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.939974 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") pod \"94531b92-db3d-4e91-acbc-ce321d2363a6\" (UID: \"94531b92-db3d-4e91-acbc-ce321d2363a6\") " Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.940783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities" (OuterVolumeSpecName: "utilities") pod "94531b92-db3d-4e91-acbc-ce321d2363a6" (UID: "94531b92-db3d-4e91-acbc-ce321d2363a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.946164 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p" (OuterVolumeSpecName: "kube-api-access-vjw7p") pod "94531b92-db3d-4e91-acbc-ce321d2363a6" (UID: "94531b92-db3d-4e91-acbc-ce321d2363a6"). InnerVolumeSpecName "kube-api-access-vjw7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:32:44 crc kubenswrapper[4907]: I0313 15:32:44.978644 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94531b92-db3d-4e91-acbc-ce321d2363a6" (UID: "94531b92-db3d-4e91-acbc-ce321d2363a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.041630 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.041669 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94531b92-db3d-4e91-acbc-ce321d2363a6-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.041680 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjw7p\" (UniqueName: \"kubernetes.io/projected/94531b92-db3d-4e91-acbc-ce321d2363a6-kube-api-access-vjw7p\") on node \"crc\" DevicePath \"\"" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423707 4907 generic.go:334] "Generic (PLEG): container finished" podID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" exitCode=0 Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423766 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182"} Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5nvvj" event={"ID":"94531b92-db3d-4e91-acbc-ce321d2363a6","Type":"ContainerDied","Data":"b9169ee22d40d10babc6f49b1bfe9a71b6b37f17774a82002ce20e9b46ed79d4"} Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423822 4907 scope.go:117] "RemoveContainer" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.423977 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5nvvj" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.444188 4907 scope.go:117] "RemoveContainer" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.462841 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.464731 4907 scope.go:117] "RemoveContainer" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.473619 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5nvvj"] Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.519425 4907 scope.go:117] "RemoveContainer" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" Mar 13 15:32:45 crc kubenswrapper[4907]: E0313 15:32:45.520232 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182\": container with ID starting with 28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182 not found: ID does not exist" containerID="28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520270 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182"} err="failed to get container status \"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182\": rpc error: code = NotFound desc = could not find container \"28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182\": container with ID starting with 28a2eb5a8c2e8b36c1aa29be80cf9445e6ac6f4b49f4d62964f9c7fe98f4f182 not found: ID does not exist" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520295 4907 scope.go:117] "RemoveContainer" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" Mar 13 15:32:45 crc kubenswrapper[4907]: E0313 15:32:45.520630 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50\": container with ID starting with 1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50 not found: ID does not exist" containerID="1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520660 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50"} err="failed to get container status \"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50\": rpc error: code = NotFound desc = could not find container \"1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50\": container with ID starting with 1f597e88c310a274ade7c6495c51734c8934b5e89f3a0a2ab10dee76116abb50 not found: ID does not exist" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.520703 4907 scope.go:117] "RemoveContainer" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" Mar 13 15:32:45 crc kubenswrapper[4907]: E0313 15:32:45.521106 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b\": container with ID starting with 1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b not found: ID does not exist" containerID="1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.521159 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b"} err="failed to get container status \"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b\": rpc error: code = NotFound desc = could not find container \"1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b\": container with ID starting with 1287840fe6e969d4fb6a3ec192005abc0045588fa52933a0bfda715eb3a10a3b not found: ID does not exist" Mar 13 15:32:45 crc kubenswrapper[4907]: I0313 15:32:45.792743 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" path="/var/lib/kubelet/pods/94531b92-db3d-4e91-acbc-ce321d2363a6/volumes" Mar 13 15:32:49 crc kubenswrapper[4907]: I0313 15:32:49.782756 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:32:49 crc kubenswrapper[4907]: E0313 15:32:49.783569 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352086 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 15:32:53 crc kubenswrapper[4907]: E0313 15:32:53.352704 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352720 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" Mar 13 15:32:53 crc kubenswrapper[4907]: E0313 15:32:53.352759 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-content" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352770 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-content" Mar 13 15:32:53 crc kubenswrapper[4907]: E0313 15:32:53.352794 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-utilities" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.352804 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="extract-utilities" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.353068 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="94531b92-db3d-4e91-acbc-ce321d2363a6" containerName="registry-server" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.353735 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.361661 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mqs4s" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.374181 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.471337 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.471393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.572767 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.572808 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.576253 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.576319 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f298a5b7b791478733605b03fd2767b8f1abec427d368f68c943bb89527cfeeb/globalmount\"" pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.595783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.606479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"mariadb-copy-data\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " pod="openstack/mariadb-copy-data" Mar 13 15:32:53 crc kubenswrapper[4907]: I0313 15:32:53.682296 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.231001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.485008 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerStarted","Data":"be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a"} Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.485057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerStarted","Data":"763a1aa9a599cf6e70f4f78d5d2da41195b4bd8f467a89e3d789f85b36605af1"} Mar 13 15:32:54 crc kubenswrapper[4907]: I0313 15:32:54.504629 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.50460878 podStartE2EDuration="2.50460878s" podCreationTimestamp="2026-03-13 15:32:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:32:54.500233161 +0000 UTC m=+5273.400020870" watchObservedRunningTime="2026-03-13 15:32:54.50460878 +0000 UTC m=+5273.404396469" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.359899 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.361670 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.367874 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.530914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"mariadb-client\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.632467 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"mariadb-client\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.660719 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"mariadb-client\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " pod="openstack/mariadb-client" Mar 13 15:32:57 crc kubenswrapper[4907]: I0313 15:32:57.714214 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.123807 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:58 crc kubenswrapper[4907]: W0313 15:32:58.126659 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ba8510f_f827_4fc5_a32b_f64153e33f5e.slice/crio-b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab WatchSource:0}: Error finding container b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab: Status 404 returned error can't find the container with id b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.525931 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerID="3c35c6992ec9b87a4cfcf6f80130463db4ea23212e37b3ceef2bd0131f1b10b1" exitCode=0 Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.526018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4ba8510f-f827-4fc5-a32b-f64153e33f5e","Type":"ContainerDied","Data":"3c35c6992ec9b87a4cfcf6f80130463db4ea23212e37b3ceef2bd0131f1b10b1"} Mar 13 15:32:58 crc kubenswrapper[4907]: I0313 15:32:58.526255 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4ba8510f-f827-4fc5-a32b-f64153e33f5e","Type":"ContainerStarted","Data":"b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab"} Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.872041 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.896843 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_4ba8510f-f827-4fc5-a32b-f64153e33f5e/mariadb-client/0.log" Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.930390 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.938041 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.963637 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") pod \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\" (UID: \"4ba8510f-f827-4fc5-a32b-f64153e33f5e\") " Mar 13 15:32:59 crc kubenswrapper[4907]: I0313 15:32:59.971806 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl" (OuterVolumeSpecName: "kube-api-access-6tkbl") pod "4ba8510f-f827-4fc5-a32b-f64153e33f5e" (UID: "4ba8510f-f827-4fc5-a32b-f64153e33f5e"). InnerVolumeSpecName "kube-api-access-6tkbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.063801 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:00 crc kubenswrapper[4907]: E0313 15:33:00.064164 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerName="mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.064185 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerName="mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.064405 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" containerName="mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.064937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.066296 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tkbl\" (UniqueName: \"kubernetes.io/projected/4ba8510f-f827-4fc5-a32b-f64153e33f5e-kube-api-access-6tkbl\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.079401 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.168048 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"mariadb-client\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.269583 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"mariadb-client\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.285445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"mariadb-client\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.388035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.541796 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b642f87690b3f00088c572d86d8921572a5e11af8fab4078310397ab7d0aaeab" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.541863 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.575237 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" podUID="6616b212-937e-4100-9007-c529da51469a" Mar 13 15:33:00 crc kubenswrapper[4907]: I0313 15:33:00.784995 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:00 crc kubenswrapper[4907]: W0313 15:33:00.787208 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6616b212_937e_4100_9007_c529da51469a.slice/crio-bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1 WatchSource:0}: Error finding container bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1: Status 404 returned error can't find the container with id bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1 Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.549580 4907 generic.go:334] "Generic (PLEG): container finished" podID="6616b212-937e-4100-9007-c529da51469a" containerID="2c2f32fc7b86119b10010b256cedf2bb75451fdf3804d5c1b2c61e62707d4b83" exitCode=0 Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.549622 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6616b212-937e-4100-9007-c529da51469a","Type":"ContainerDied","Data":"2c2f32fc7b86119b10010b256cedf2bb75451fdf3804d5c1b2c61e62707d4b83"} Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.549655 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"6616b212-937e-4100-9007-c529da51469a","Type":"ContainerStarted","Data":"bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1"} Mar 13 15:33:01 crc kubenswrapper[4907]: I0313 15:33:01.793149 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba8510f-f827-4fc5-a32b-f64153e33f5e" path="/var/lib/kubelet/pods/4ba8510f-f827-4fc5-a32b-f64153e33f5e/volumes" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.784092 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:02 crc kubenswrapper[4907]: E0313 15:33:02.785207 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.864427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.881518 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_6616b212-937e-4100-9007-c529da51469a/mariadb-client/0.log" Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.906075 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.909529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") pod \"6616b212-937e-4100-9007-c529da51469a\" (UID: \"6616b212-937e-4100-9007-c529da51469a\") " Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.914318 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Mar 13 15:33:02 crc kubenswrapper[4907]: I0313 15:33:02.916220 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c" (OuterVolumeSpecName: "kube-api-access-8f58c") pod "6616b212-937e-4100-9007-c529da51469a" (UID: "6616b212-937e-4100-9007-c529da51469a"). InnerVolumeSpecName "kube-api-access-8f58c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.010607 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f58c\" (UniqueName: \"kubernetes.io/projected/6616b212-937e-4100-9007-c529da51469a-kube-api-access-8f58c\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.571233 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbea2db874a2a8327fe7e7e46748c830df370fb5bb85720bc7d40582c81c4bb1" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.571539 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Mar 13 15:33:03 crc kubenswrapper[4907]: I0313 15:33:03.792984 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6616b212-937e-4100-9007-c529da51469a" path="/var/lib/kubelet/pods/6616b212-937e-4100-9007-c529da51469a/volumes" Mar 13 15:33:13 crc kubenswrapper[4907]: I0313 15:33:13.782966 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:13 crc kubenswrapper[4907]: E0313 15:33:13.783720 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:26 crc kubenswrapper[4907]: I0313 15:33:26.783291 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:26 crc kubenswrapper[4907]: E0313 15:33:26.786875 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:27 crc kubenswrapper[4907]: I0313 15:33:27.835940 4907 scope.go:117] "RemoveContainer" containerID="83c101cd6c30ca361c7f00d6a637893118fddd69105f31e3b1c72e730b52960a" Mar 13 15:33:27 crc kubenswrapper[4907]: I0313 15:33:27.856287 4907 scope.go:117] "RemoveContainer" containerID="8f8264e04faa67742d13e20b2a8a6a04d3b6912538ef54a5202a7e841e98d29e" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.133323 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: E0313 15:33:34.134634 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6616b212-937e-4100-9007-c529da51469a" containerName="mariadb-client" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.134653 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6616b212-937e-4100-9007-c529da51469a" containerName="mariadb-client" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.134955 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6616b212-937e-4100-9007-c529da51469a" containerName="mariadb-client" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.135767 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.138259 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.138289 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-s57jd" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.138464 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.154131 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.155894 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.161718 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.167581 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.200534 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.210598 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.225290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270469 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270561 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdlpw\" (UniqueName: \"kubernetes.io/projected/e5ec94c1-f60b-4156-b3ba-278db827c9b8-kube-api-access-cdlpw\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270619 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw2xr\" (UniqueName: \"kubernetes.io/projected/b188658f-1e5c-49d4-be19-0a888caddb43-kube-api-access-xw2xr\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b188658f-1e5c-49d4-be19-0a888caddb43-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bffaea76-e210-44d1-8caa-ebf82edb8ef0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-config\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270697 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5ec94c1-f60b-4156-b3ba-278db827c9b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270721 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec94c1-f60b-4156-b3ba-278db827c9b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270834 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.270995 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271182 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271270 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b188658f-1e5c-49d4-be19-0a888caddb43-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271349 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bffaea76-e210-44d1-8caa-ebf82edb8ef0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shc8z\" (UniqueName: \"kubernetes.io/projected/bffaea76-e210-44d1-8caa-ebf82edb8ef0-kube-api-access-shc8z\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271516 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.271635 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.343111 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.345050 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.347709 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tpm7x" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.347718 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.352660 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.355319 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.372424 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.383120 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390177 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390572 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b188658f-1e5c-49d4-be19-0a888caddb43-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390653 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bffaea76-e210-44d1-8caa-ebf82edb8ef0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390794 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shc8z\" (UniqueName: \"kubernetes.io/projected/bffaea76-e210-44d1-8caa-ebf82edb8ef0-kube-api-access-shc8z\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.390977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391032 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391136 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdlpw\" (UniqueName: \"kubernetes.io/projected/e5ec94c1-f60b-4156-b3ba-278db827c9b8-kube-api-access-cdlpw\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391210 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw2xr\" (UniqueName: \"kubernetes.io/projected/b188658f-1e5c-49d4-be19-0a888caddb43-kube-api-access-xw2xr\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391220 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b188658f-1e5c-49d4-be19-0a888caddb43-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b188658f-1e5c-49d4-be19-0a888caddb43-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391279 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bffaea76-e210-44d1-8caa-ebf82edb8ef0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391307 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-config\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5ec94c1-f60b-4156-b3ba-278db827c9b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391362 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec94c1-f60b-4156-b3ba-278db827c9b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391386 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391389 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bffaea76-e210-44d1-8caa-ebf82edb8ef0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.391420 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.392004 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-config\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.396523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-config\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.406762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec94c1-f60b-4156-b3ba-278db827c9b8-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.409503 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.411562 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b188658f-1e5c-49d4-be19-0a888caddb43-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.411607 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.411636 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8198106df7b96a617bc2acc54b8057faf23b369987a114c32f994974a6cdd7ae/globalmount\"" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.412453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-config\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.413641 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e5ec94c1-f60b-4156-b3ba-278db827c9b8-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.413656 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bffaea76-e210-44d1-8caa-ebf82edb8ef0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.413895 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e5ec94c1-f60b-4156-b3ba-278db827c9b8-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.415158 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.415359 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/22668b1ed8c08d70322619ed0b00585c03f06e051ce689e13cc883c7d890ed36/globalmount\"" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.417553 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bffaea76-e210-44d1-8caa-ebf82edb8ef0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.419570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw2xr\" (UniqueName: \"kubernetes.io/projected/b188658f-1e5c-49d4-be19-0a888caddb43-kube-api-access-xw2xr\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.424540 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b188658f-1e5c-49d4-be19-0a888caddb43-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.432314 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.434995 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shc8z\" (UniqueName: \"kubernetes.io/projected/bffaea76-e210-44d1-8caa-ebf82edb8ef0-kube-api-access-shc8z\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.437792 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdlpw\" (UniqueName: \"kubernetes.io/projected/e5ec94c1-f60b-4156-b3ba-278db827c9b8-kube-api-access-cdlpw\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.440920 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.440953 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/00f7424f6f9b73ded1a9336a9e06d79a0db8f52c63e13ea4f24769687d2a94b2/globalmount\"" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.449950 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.460938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3bdb22ed-1952-4cca-8e2a-e4c4e2597e8c\") pod \"ovsdbserver-sb-1\" (UID: \"e5ec94c1-f60b-4156-b3ba-278db827c9b8\") " pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.465272 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4247ec19-7024-4129-80de-23a6ba6ce376\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4247ec19-7024-4129-80de-23a6ba6ce376\") pod \"ovsdbserver-sb-0\" (UID: \"bffaea76-e210-44d1-8caa-ebf82edb8ef0\") " pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.468146 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8e01f892-be99-49f8-b8aa-bf80a68ae42b\") pod \"ovsdbserver-sb-2\" (UID: \"b188658f-1e5c-49d4-be19-0a888caddb43\") " pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.478522 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167eb31-6b82-44f9-bf66-465b64a91226-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494921 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.494995 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfvnc\" (UniqueName: \"kubernetes.io/projected/a167eb31-6b82-44f9-bf66-465b64a91226-kube-api-access-qfvnc\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495034 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45mq2\" (UniqueName: \"kubernetes.io/projected/3ca7bf8f-4514-493e-9385-75cfcd0807f9-kube-api-access-45mq2\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495078 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495116 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-config\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495153 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-config\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495209 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bf8f-4514-493e-9385-75cfcd0807f9-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495317 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a167eb31-6b82-44f9-bf66-465b64a91226-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.495931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bf8f-4514-493e-9385-75cfcd0807f9-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.496014 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.496147 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.497102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-config\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.497855 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.497921 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbbbq\" (UniqueName: \"kubernetes.io/projected/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-kube-api-access-lbbbq\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.506544 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.599151 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167eb31-6b82-44f9-bf66-465b64a91226-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600451 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600470 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600493 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfvnc\" (UniqueName: \"kubernetes.io/projected/a167eb31-6b82-44f9-bf66-465b64a91226-kube-api-access-qfvnc\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45mq2\" (UniqueName: \"kubernetes.io/projected/3ca7bf8f-4514-493e-9385-75cfcd0807f9-kube-api-access-45mq2\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600569 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-config\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-config\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600617 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bf8f-4514-493e-9385-75cfcd0807f9-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a167eb31-6b82-44f9-bf66-465b64a91226-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600703 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600730 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bf8f-4514-493e-9385-75cfcd0807f9-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-config\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600911 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.600946 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbbbq\" (UniqueName: \"kubernetes.io/projected/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-kube-api-access-lbbbq\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.604558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.604857 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.605197 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-config\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.605223 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ca7bf8f-4514-493e-9385-75cfcd0807f9-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.607453 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ca7bf8f-4514-493e-9385-75cfcd0807f9-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.608078 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-config\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.609675 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ca7bf8f-4514-493e-9385-75cfcd0807f9-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.609681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.610190 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.610219 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/554de8a28393bb5481936a1456b52e6abac7129d07afc1025c887f49696ce35e/globalmount\"" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.611012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-config\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.611231 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a167eb31-6b82-44f9-bf66-465b64a91226-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.611551 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a167eb31-6b82-44f9-bf66-465b64a91226-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614034 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614076 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/99ba545b23bc3775acd4c535d721325e522f269a911518ef755286a88dbac8f2/globalmount\"" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614192 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.614242 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7bde764fd3ca106780c183d470e3dca511bd76d320905b722bf14f86ea815a69/globalmount\"" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.624588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a167eb31-6b82-44f9-bf66-465b64a91226-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.624963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfvnc\" (UniqueName: \"kubernetes.io/projected/a167eb31-6b82-44f9-bf66-465b64a91226-kube-api-access-qfvnc\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.626502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbbbq\" (UniqueName: \"kubernetes.io/projected/7e270de1-912f-48f9-bfc7-e415fa5c7b5d-kube-api-access-lbbbq\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.629604 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45mq2\" (UniqueName: \"kubernetes.io/projected/3ca7bf8f-4514-493e-9385-75cfcd0807f9-kube-api-access-45mq2\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.661215 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f5e5ec9c-2696-40da-aa58-a0cdbb969cd9\") pod \"ovsdbserver-nb-0\" (UID: \"a167eb31-6b82-44f9-bf66-465b64a91226\") " pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.668829 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b7e3d76-850f-42aa-9232-939baf06bab0\") pod \"ovsdbserver-nb-1\" (UID: \"7e270de1-912f-48f9-bfc7-e415fa5c7b5d\") " pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.674257 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-be1fde5e-0257-4035-803c-6c276dcd538c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-be1fde5e-0257-4035-803c-6c276dcd538c\") pod \"ovsdbserver-nb-2\" (UID: \"3ca7bf8f-4514-493e-9385-75cfcd0807f9\") " pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.858055 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.872320 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:34 crc kubenswrapper[4907]: I0313 15:33:34.963517 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.017494 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 15:33:35 crc kubenswrapper[4907]: W0313 15:33:35.124954 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb188658f_1e5c_49d4_be19_0a888caddb43.slice/crio-674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba WatchSource:0}: Error finding container 674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba: Status 404 returned error can't find the container with id 674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.130507 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.206445 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.418136 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.529205 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Mar 13 15:33:35 crc kubenswrapper[4907]: W0313 15:33:35.535802 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ca7bf8f_4514_493e_9385_75cfcd0807f9.slice/crio-3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef WatchSource:0}: Error finding container 3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef: Status 404 returned error can't find the container with id 3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef Mar 13 15:33:35 crc kubenswrapper[4907]: W0313 15:33:35.610076 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda167eb31_6b82_44f9_bf66_465b64a91226.slice/crio-79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037 WatchSource:0}: Error finding container 79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037: Status 404 returned error can't find the container with id 79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037 Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.612099 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.816992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5ec94c1-f60b-4156-b3ba-278db827c9b8","Type":"ContainerStarted","Data":"8f22dd4ec76846c8d2a37b6af861b79cc8fcee078f77e42f7ce5d93872a72cf3"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.817035 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5ec94c1-f60b-4156-b3ba-278db827c9b8","Type":"ContainerStarted","Data":"2c3fa3c8d0e82f5efce1785d752c20bdf28c4cd1e70d951f529de9c7869a9e3b"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.817045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"e5ec94c1-f60b-4156-b3ba-278db827c9b8","Type":"ContainerStarted","Data":"0bff93d9d737cb5a76a3af81265b6b53584369519d92c7569bfd52f6414fe9d3"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.820950 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a167eb31-6b82-44f9-bf66-465b64a91226","Type":"ContainerStarted","Data":"4ba727733c83b392d0aa81d74209df8981f0e659132779ee0ceb6a05ef7d5103"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.821002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a167eb31-6b82-44f9-bf66-465b64a91226","Type":"ContainerStarted","Data":"79ebf4c6ba715cb4043d5908d97a193e951453a828adb94d17d744b98922a037"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.828177 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bffaea76-e210-44d1-8caa-ebf82edb8ef0","Type":"ContainerStarted","Data":"90dadbe657fd114fe2e914afa903244ef949d9111d57f162f668daadb25f1921"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.828233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bffaea76-e210-44d1-8caa-ebf82edb8ef0","Type":"ContainerStarted","Data":"d4d1ad2a84a27593b08932e7a490f0cb17eb191199bc9374857ed81b7661c44e"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.828245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bffaea76-e210-44d1-8caa-ebf82edb8ef0","Type":"ContainerStarted","Data":"59294b2cb6419dec7fe1f1dbada2c9df28cf8f11feb23127830df996d8ed5535"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.842955 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b188658f-1e5c-49d4-be19-0a888caddb43","Type":"ContainerStarted","Data":"0670fcbcd8be942fc9083ff1c66a6253e07eaec25465a3bf14ed1e0c14798409"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.843001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b188658f-1e5c-49d4-be19-0a888caddb43","Type":"ContainerStarted","Data":"aef67d6a1ec2fcb17e8dcb7f26ca4c158448ccf469b72e5cca1a5e99b7cd2d21"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.843013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b188658f-1e5c-49d4-be19-0a888caddb43","Type":"ContainerStarted","Data":"674402c7ad6b0314499d0c1c017cc7ebeb988958ee9224c14b3b9c8ef682dfba"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.844956 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=2.844935366 podStartE2EDuration="2.844935366s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:35.837669516 +0000 UTC m=+5314.737457225" watchObservedRunningTime="2026-03-13 15:33:35.844935366 +0000 UTC m=+5314.744723065" Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.845907 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"7e270de1-912f-48f9-bfc7-e415fa5c7b5d","Type":"ContainerStarted","Data":"d7819948441d190c863d4f57f53dd980b3bbb30335e7487d4366124beb0e6c1d"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.845947 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"7e270de1-912f-48f9-bfc7-e415fa5c7b5d","Type":"ContainerStarted","Data":"147d938f69e2e32e5d6fee7c068767956ae44d04924674daefa96fdf56a2a532"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.851057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3ca7bf8f-4514-493e-9385-75cfcd0807f9","Type":"ContainerStarted","Data":"8404c331259c77750125e3e64294e712b997d439ae3bb0cde654a622897811bd"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.851095 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3ca7bf8f-4514-493e-9385-75cfcd0807f9","Type":"ContainerStarted","Data":"3b780eb2bb9a20e04ad7d5ac0beab1e2e4343780f0a2d6530af68108f6b4beef"} Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.861496 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.861480548 podStartE2EDuration="2.861480548s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:35.855303799 +0000 UTC m=+5314.755091488" watchObservedRunningTime="2026-03-13 15:33:35.861480548 +0000 UTC m=+5314.761268237" Mar 13 15:33:35 crc kubenswrapper[4907]: I0313 15:33:35.884217 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=2.88419758 podStartE2EDuration="2.88419758s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:35.873859387 +0000 UTC m=+5314.773647066" watchObservedRunningTime="2026-03-13 15:33:35.88419758 +0000 UTC m=+5314.783985269" Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.858582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"7e270de1-912f-48f9-bfc7-e415fa5c7b5d","Type":"ContainerStarted","Data":"e04df1932d1e43003ccafc8b156c0365035744d253a0b242b932fc2d50b2a961"} Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.861556 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"3ca7bf8f-4514-493e-9385-75cfcd0807f9","Type":"ContainerStarted","Data":"953ee43d80e6b216eca1936519f349ce31b34a7eaa34a251834139aa0548d049"} Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.863997 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a167eb31-6b82-44f9-bf66-465b64a91226","Type":"ContainerStarted","Data":"1fda18c9bb6492a6795eff525886132c05c148df97ad8119fd84e436dbecec7e"} Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.877237 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.87720873 podStartE2EDuration="3.87720873s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:36.877112027 +0000 UTC m=+5315.776899716" watchObservedRunningTime="2026-03-13 15:33:36.87720873 +0000 UTC m=+5315.776996419" Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.899251 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.899232342 podStartE2EDuration="3.899232342s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:36.893567717 +0000 UTC m=+5315.793355426" watchObservedRunningTime="2026-03-13 15:33:36.899232342 +0000 UTC m=+5315.799020031" Mar 13 15:33:36 crc kubenswrapper[4907]: I0313 15:33:36.918501 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.918443188 podStartE2EDuration="3.918443188s" podCreationTimestamp="2026-03-13 15:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:36.908859876 +0000 UTC m=+5315.808647585" watchObservedRunningTime="2026-03-13 15:33:36.918443188 +0000 UTC m=+5315.818230877" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.479045 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.496774 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.506971 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.859084 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.873626 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:37 crc kubenswrapper[4907]: I0313 15:33:37.963951 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.478805 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.497390 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.507128 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.859101 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.873832 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:39 crc kubenswrapper[4907]: I0313 15:33:39.963988 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.535759 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.542773 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.555656 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.584867 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.598108 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.598794 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.783477 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:40 crc kubenswrapper[4907]: E0313 15:33:40.785302 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.843485 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.845412 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.847506 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.852616 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.897595 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.914375 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.952899 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.960723 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966725 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966775 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:40 crc kubenswrapper[4907]: I0313 15:33:40.966969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.006006 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.047854 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.067930 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.068010 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.068863 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.068863 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.069047 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.069090 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.069745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.086140 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"dnsmasq-dns-794d85d9cf-xl4f8\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.168655 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.369465 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.408798 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.419310 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.424764 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.448326 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476522 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476825 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.476872 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578125 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578220 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578270 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.578998 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.579176 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.579344 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.579627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.596987 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"dnsmasq-dns-65b7b7b995-wl96j\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.665181 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.746528 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.924240 4907 generic.go:334] "Generic (PLEG): container finished" podID="72846279-1333-4762-b5c9-cd7b25492843" containerID="b312e10aee1a09c0f414d5ab55a14272c1683d9e1cf46fbb198ff6477ba0ca4f" exitCode=0 Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.924392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" event={"ID":"72846279-1333-4762-b5c9-cd7b25492843","Type":"ContainerDied","Data":"b312e10aee1a09c0f414d5ab55a14272c1683d9e1cf46fbb198ff6477ba0ca4f"} Mar 13 15:33:41 crc kubenswrapper[4907]: I0313 15:33:41.924657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" event={"ID":"72846279-1333-4762-b5c9-cd7b25492843","Type":"ContainerStarted","Data":"847c3ddce6826b442d1b5ec210f83bfadfb60c76eab28f422fd1cc1f341d21d1"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.210763 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.212384 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:33:42 crc kubenswrapper[4907]: W0313 15:33:42.217069 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e01f82_fb82_4351_a637_3b885e663e0e.slice/crio-2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101 WatchSource:0}: Error finding container 2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101: Status 404 returned error can't find the container with id 2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101 Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.388890 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.388954 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.389331 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.389386 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") pod \"72846279-1333-4762-b5c9-cd7b25492843\" (UID: \"72846279-1333-4762-b5c9-cd7b25492843\") " Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.396374 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp" (OuterVolumeSpecName: "kube-api-access-6jdqp") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "kube-api-access-6jdqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.412176 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config" (OuterVolumeSpecName: "config") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.420654 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.420767 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "72846279-1333-4762-b5c9-cd7b25492843" (UID: "72846279-1333-4762-b5c9-cd7b25492843"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492070 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492407 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jdqp\" (UniqueName: \"kubernetes.io/projected/72846279-1333-4762-b5c9-cd7b25492843-kube-api-access-6jdqp\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492478 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.492539 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72846279-1333-4762-b5c9-cd7b25492843-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.933568 4907 generic.go:334] "Generic (PLEG): container finished" podID="49e01f82-fb82-4351-a637-3b885e663e0e" containerID="7829d0510750d5cae8df3422b54b042bd33e40f132a8e3e16f6eec0a8952963b" exitCode=0 Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.933638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerDied","Data":"7829d0510750d5cae8df3422b54b042bd33e40f132a8e3e16f6eec0a8952963b"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.933666 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerStarted","Data":"2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.935139 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" event={"ID":"72846279-1333-4762-b5c9-cd7b25492843","Type":"ContainerDied","Data":"847c3ddce6826b442d1b5ec210f83bfadfb60c76eab28f422fd1cc1f341d21d1"} Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.935178 4907 scope.go:117] "RemoveContainer" containerID="b312e10aee1a09c0f414d5ab55a14272c1683d9e1cf46fbb198ff6477ba0ca4f" Mar 13 15:33:42 crc kubenswrapper[4907]: I0313 15:33:42.935186 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-794d85d9cf-xl4f8" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.006986 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.022343 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-794d85d9cf-xl4f8"] Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.795158 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72846279-1333-4762-b5c9-cd7b25492843" path="/var/lib/kubelet/pods/72846279-1333-4762-b5c9-cd7b25492843/volumes" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.916598 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Mar 13 15:33:43 crc kubenswrapper[4907]: E0313 15:33:43.916936 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72846279-1333-4762-b5c9-cd7b25492843" containerName="init" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.916952 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="72846279-1333-4762-b5c9-cd7b25492843" containerName="init" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.917292 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="72846279-1333-4762-b5c9-cd7b25492843" containerName="init" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.918543 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.921165 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.926942 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.952182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerStarted","Data":"a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965"} Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.952321 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:43 crc kubenswrapper[4907]: I0313 15:33:43.975163 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" podStartSLOduration=2.975146467 podStartE2EDuration="2.975146467s" podCreationTimestamp="2026-03-13 15:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:43.969126812 +0000 UTC m=+5322.868914551" watchObservedRunningTime="2026-03-13 15:33:43.975146467 +0000 UTC m=+5322.874934156" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.014539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.014589 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.014659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.117091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.117234 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.117268 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.121535 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.122308 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.122344 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5d7cb02ae805cd0ef340a8d0e965e05c32a025a0afcab88074ddc006f8992ebd/globalmount\"" pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.140377 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.149971 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"ovn-copy-data\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.238162 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.709764 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 15:33:44 crc kubenswrapper[4907]: W0313 15:33:44.717765 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0c0f09a_8646_4326_86e8_5bd9123d78ca.slice/crio-afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be WatchSource:0}: Error finding container afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be: Status 404 returned error can't find the container with id afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be Mar 13 15:33:44 crc kubenswrapper[4907]: I0313 15:33:44.961388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerStarted","Data":"afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be"} Mar 13 15:33:45 crc kubenswrapper[4907]: I0313 15:33:45.973577 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerStarted","Data":"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c"} Mar 13 15:33:45 crc kubenswrapper[4907]: I0313 15:33:45.994600 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.45972068 podStartE2EDuration="3.99457321s" podCreationTimestamp="2026-03-13 15:33:42 +0000 UTC" firstStartedPulling="2026-03-13 15:33:44.720876628 +0000 UTC m=+5323.620664317" lastFinishedPulling="2026-03-13 15:33:45.255729158 +0000 UTC m=+5324.155516847" observedRunningTime="2026-03-13 15:33:45.993075579 +0000 UTC m=+5324.892863288" watchObservedRunningTime="2026-03-13 15:33:45.99457321 +0000 UTC m=+5324.894360929" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.058444 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.073689 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.079756 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.082647 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-rgmnm" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.082708 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.082871 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf6kz\" (UniqueName: \"kubernetes.io/projected/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-kube-api-access-cf6kz\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134634 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134653 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-config\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.134772 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-scripts\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236289 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf6kz\" (UniqueName: \"kubernetes.io/projected/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-kube-api-access-cf6kz\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236382 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236410 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-config\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236472 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-scripts\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.236990 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.237358 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-scripts\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.237485 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-config\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.242589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.253025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf6kz\" (UniqueName: \"kubernetes.io/projected/1b95d939-59f1-4ba9-b39a-eb255ab7aa77-kube-api-access-cf6kz\") pod \"ovn-northd-0\" (UID: \"1b95d939-59f1-4ba9-b39a-eb255ab7aa77\") " pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.405937 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.748126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.849792 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.850024 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" containerID="cri-o://1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2" gracePeriod=10 Mar 13 15:33:51 crc kubenswrapper[4907]: I0313 15:33:51.890633 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 15:33:51 crc kubenswrapper[4907]: W0313 15:33:51.944474 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b95d939_59f1_4ba9_b39a_eb255ab7aa77.slice/crio-f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4 WatchSource:0}: Error finding container f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4: Status 404 returned error can't find the container with id f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4 Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.024711 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b95d939-59f1-4ba9-b39a-eb255ab7aa77","Type":"ContainerStarted","Data":"f1360a96df510c54ad081cf7e101eb797e728625968b75461bc0e672bb69bda4"} Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.026413 4907 generic.go:334] "Generic (PLEG): container finished" podID="6c367cd2-6ece-429d-a930-c1ec11175248" containerID="1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2" exitCode=0 Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.026449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerDied","Data":"1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2"} Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.252291 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.368616 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") pod \"6c367cd2-6ece-429d-a930-c1ec11175248\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.368703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") pod \"6c367cd2-6ece-429d-a930-c1ec11175248\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.368803 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") pod \"6c367cd2-6ece-429d-a930-c1ec11175248\" (UID: \"6c367cd2-6ece-429d-a930-c1ec11175248\") " Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.373478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p" (OuterVolumeSpecName: "kube-api-access-rhz6p") pod "6c367cd2-6ece-429d-a930-c1ec11175248" (UID: "6c367cd2-6ece-429d-a930-c1ec11175248"). InnerVolumeSpecName "kube-api-access-rhz6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.401621 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config" (OuterVolumeSpecName: "config") pod "6c367cd2-6ece-429d-a930-c1ec11175248" (UID: "6c367cd2-6ece-429d-a930-c1ec11175248"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.406746 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c367cd2-6ece-429d-a930-c1ec11175248" (UID: "6c367cd2-6ece-429d-a930-c1ec11175248"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.470277 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhz6p\" (UniqueName: \"kubernetes.io/projected/6c367cd2-6ece-429d-a930-c1ec11175248-kube-api-access-rhz6p\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.470306 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:52 crc kubenswrapper[4907]: I0313 15:33:52.470315 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c367cd2-6ece-429d-a930-c1ec11175248-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.042419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" event={"ID":"6c367cd2-6ece-429d-a930-c1ec11175248","Type":"ContainerDied","Data":"bc53c9c5cde7280652faad4d880d4ee4a2f7625ff891a79d0f242d66a55b312a"} Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.042474 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-684c864bc9-p4j95" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.042482 4907 scope.go:117] "RemoveContainer" containerID="1b6e2c3bc8a46bfa72edd14d9923048107755f57239082f4595ce1f2d9b0f3a2" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.045529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b95d939-59f1-4ba9-b39a-eb255ab7aa77","Type":"ContainerStarted","Data":"4e605736636de74c497de794d5fba08d765bee51520f0a7d4c6c7d3da9d0341f"} Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.045562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1b95d939-59f1-4ba9-b39a-eb255ab7aa77","Type":"ContainerStarted","Data":"1b1a680cf63cb8b00e032e250e346fc2a5d3bfc5566f746214754e555cb13ce5"} Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.046353 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.060408 4907 scope.go:117] "RemoveContainer" containerID="dfbeef34d4d9c1c564438050b366f2c5cc11365a8117cb19955310b23bc22e05" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.070725 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.07070575 podStartE2EDuration="2.07070575s" podCreationTimestamp="2026-03-13 15:33:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:33:53.063276047 +0000 UTC m=+5331.963063776" watchObservedRunningTime="2026-03-13 15:33:53.07070575 +0000 UTC m=+5331.970493439" Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.107293 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.119532 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-684c864bc9-p4j95"] Mar 13 15:33:53 crc kubenswrapper[4907]: I0313 15:33:53.790959 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" path="/var/lib/kubelet/pods/6c367cd2-6ece-429d-a930-c1ec11175248/volumes" Mar 13 15:33:55 crc kubenswrapper[4907]: I0313 15:33:55.782956 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:33:55 crc kubenswrapper[4907]: E0313 15:33:55.783422 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.652691 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:33:56 crc kubenswrapper[4907]: E0313 15:33:56.653405 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="init" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.653434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="init" Mar 13 15:33:56 crc kubenswrapper[4907]: E0313 15:33:56.653463 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.653472 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.653690 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c367cd2-6ece-429d-a930-c1ec11175248" containerName="dnsmasq-dns" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.654306 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.663960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.739683 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.739863 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.763130 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.764425 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.771295 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.773484 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841177 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841295 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.841995 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.868356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"keystone-db-create-gxjv5\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.942716 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.942756 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.943411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.960392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"keystone-48f2-account-create-update-wvks8\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:56 crc kubenswrapper[4907]: I0313 15:33:56.969174 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:33:57 crc kubenswrapper[4907]: I0313 15:33:57.081956 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:33:58 crc kubenswrapper[4907]: I0313 15:33:58.022447 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:33:58 crc kubenswrapper[4907]: I0313 15:33:58.089819 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gxjv5" event={"ID":"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058","Type":"ContainerStarted","Data":"ad6e40e290e5e7fd73118a9466156d34be11e432c2810a7a56144057ee355e41"} Mar 13 15:33:58 crc kubenswrapper[4907]: I0313 15:33:58.114409 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:33:58 crc kubenswrapper[4907]: W0313 15:33:58.128475 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb03a1328_cfc5_4ff6_88b9_6523f4378708.slice/crio-82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b WatchSource:0}: Error finding container 82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b: Status 404 returned error can't find the container with id 82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.100341 4907 generic.go:334] "Generic (PLEG): container finished" podID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerID="d95988efdbf9f67c502e797aeeb6493599a1bcc1156251e60584683248ab36b0" exitCode=0 Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.100441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48f2-account-create-update-wvks8" event={"ID":"b03a1328-cfc5-4ff6-88b9-6523f4378708","Type":"ContainerDied","Data":"d95988efdbf9f67c502e797aeeb6493599a1bcc1156251e60584683248ab36b0"} Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.100721 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48f2-account-create-update-wvks8" event={"ID":"b03a1328-cfc5-4ff6-88b9-6523f4378708","Type":"ContainerStarted","Data":"82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b"} Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.104241 4907 generic.go:334] "Generic (PLEG): container finished" podID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerID="6f94031107f6bbf3f09d35d19fba9b0dfe08f6a055661f1e04241d246d42472d" exitCode=0 Mar 13 15:33:59 crc kubenswrapper[4907]: I0313 15:33:59.104288 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gxjv5" event={"ID":"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058","Type":"ContainerDied","Data":"6f94031107f6bbf3f09d35d19fba9b0dfe08f6a055661f1e04241d246d42472d"} Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.139819 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.141326 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.143677 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.143870 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.144067 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.176033 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.202315 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"auto-csr-approver-29556934-wf7zr\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.303937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"auto-csr-approver-29556934-wf7zr\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.333570 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"auto-csr-approver-29556934-wf7zr\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.469066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.569098 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.573533 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710101 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") pod \"b03a1328-cfc5-4ff6-88b9-6523f4378708\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710158 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") pod \"b03a1328-cfc5-4ff6-88b9-6523f4378708\" (UID: \"b03a1328-cfc5-4ff6-88b9-6523f4378708\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") pod \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710351 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") pod \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\" (UID: \"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058\") " Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.710957 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" (UID: "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.711487 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b03a1328-cfc5-4ff6-88b9-6523f4378708" (UID: "b03a1328-cfc5-4ff6-88b9-6523f4378708"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.714545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz" (OuterVolumeSpecName: "kube-api-access-979dz") pod "b03a1328-cfc5-4ff6-88b9-6523f4378708" (UID: "b03a1328-cfc5-4ff6-88b9-6523f4378708"). InnerVolumeSpecName "kube-api-access-979dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.714622 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8" (OuterVolumeSpecName: "kube-api-access-q45f8") pod "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" (UID: "9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058"). InnerVolumeSpecName "kube-api-access-q45f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812259 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03a1328-cfc5-4ff6-88b9-6523f4378708-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812300 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-979dz\" (UniqueName: \"kubernetes.io/projected/b03a1328-cfc5-4ff6-88b9-6523f4378708-kube-api-access-979dz\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812313 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.812322 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q45f8\" (UniqueName: \"kubernetes.io/projected/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058-kube-api-access-q45f8\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:00 crc kubenswrapper[4907]: I0313 15:34:00.897778 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:34:00 crc kubenswrapper[4907]: W0313 15:34:00.899640 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod041cab28_d0e1_4ec9_a0cb_92176be369aa.slice/crio-1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c WatchSource:0}: Error finding container 1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c: Status 404 returned error can't find the container with id 1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.121764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-48f2-account-create-update-wvks8" event={"ID":"b03a1328-cfc5-4ff6-88b9-6523f4378708","Type":"ContainerDied","Data":"82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b"} Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.121804 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-48f2-account-create-update-wvks8" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.121815 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82735ca0c85dbbd92ff63172f6177cf41836602cab104a6008e63c9d13624f5b" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.126933 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gxjv5" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.126980 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gxjv5" event={"ID":"9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058","Type":"ContainerDied","Data":"ad6e40e290e5e7fd73118a9466156d34be11e432c2810a7a56144057ee355e41"} Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.127122 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad6e40e290e5e7fd73118a9466156d34be11e432c2810a7a56144057ee355e41" Mar 13 15:34:01 crc kubenswrapper[4907]: I0313 15:34:01.128204 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" event={"ID":"041cab28-d0e1-4ec9-a0cb-92176be369aa","Type":"ContainerStarted","Data":"1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c"} Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.389376 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:34:02 crc kubenswrapper[4907]: E0313 15:34:02.390366 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerName="mariadb-account-create-update" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390386 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerName="mariadb-account-create-update" Mar 13 15:34:02 crc kubenswrapper[4907]: E0313 15:34:02.390429 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerName="mariadb-database-create" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390438 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerName="mariadb-database-create" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390828 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" containerName="mariadb-account-create-update" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.390862 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" containerName="mariadb-database-create" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.392307 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.394580 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.394810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.395059 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.395399 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.400529 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.442845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.442943 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.443081 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.544670 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.544729 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.544762 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.558649 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.565843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.568177 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"keystone-db-sync-4gtrw\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:02 crc kubenswrapper[4907]: I0313 15:34:02.718379 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:03 crc kubenswrapper[4907]: I0313 15:34:03.143867 4907 generic.go:334] "Generic (PLEG): container finished" podID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerID="1eda03ad73ea12900c383cf8f9715d8d1c4690e9fdc026cd78850ed84ad83b37" exitCode=0 Mar 13 15:34:03 crc kubenswrapper[4907]: I0313 15:34:03.143975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" event={"ID":"041cab28-d0e1-4ec9-a0cb-92176be369aa","Type":"ContainerDied","Data":"1eda03ad73ea12900c383cf8f9715d8d1c4690e9fdc026cd78850ed84ad83b37"} Mar 13 15:34:03 crc kubenswrapper[4907]: I0313 15:34:03.170065 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.163457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerStarted","Data":"a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83"} Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.163763 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerStarted","Data":"3c42a8c243c0093bfc9f1f8cf6bfd418cc51eaed61e0ec45a9fbb1802050c2c9"} Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.526136 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.542094 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-4gtrw" podStartSLOduration=2.542073573 podStartE2EDuration="2.542073573s" podCreationTimestamp="2026-03-13 15:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:04.18279993 +0000 UTC m=+5343.082587619" watchObservedRunningTime="2026-03-13 15:34:04.542073573 +0000 UTC m=+5343.441861262" Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.575061 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") pod \"041cab28-d0e1-4ec9-a0cb-92176be369aa\" (UID: \"041cab28-d0e1-4ec9-a0cb-92176be369aa\") " Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.583520 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2" (OuterVolumeSpecName: "kube-api-access-kwpz2") pod "041cab28-d0e1-4ec9-a0cb-92176be369aa" (UID: "041cab28-d0e1-4ec9-a0cb-92176be369aa"). InnerVolumeSpecName "kube-api-access-kwpz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:04 crc kubenswrapper[4907]: I0313 15:34:04.677337 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwpz2\" (UniqueName: \"kubernetes.io/projected/041cab28-d0e1-4ec9-a0cb-92176be369aa-kube-api-access-kwpz2\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.172275 4907 generic.go:334] "Generic (PLEG): container finished" podID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerID="a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83" exitCode=0 Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.172329 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerDied","Data":"a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83"} Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.174458 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" event={"ID":"041cab28-d0e1-4ec9-a0cb-92176be369aa","Type":"ContainerDied","Data":"1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c"} Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.174488 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1deb6d11c2b99bf690d558f4ecc15f659c14ba85d596c123916d00f2b8567d4c" Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.174540 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556934-wf7zr" Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.591523 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.598946 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556928-nh6lw"] Mar 13 15:34:05 crc kubenswrapper[4907]: I0313 15:34:05.790793 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="566ddaa7-ea07-421b-a181-0aa768454ba6" path="/var/lib/kubelet/pods/566ddaa7-ea07-421b-a181-0aa768454ba6/volumes" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.588896 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.712795 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") pod \"f23a73ee-57ff-4d58-8812-b72f624b7739\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.712854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") pod \"f23a73ee-57ff-4d58-8812-b72f624b7739\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.713014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") pod \"f23a73ee-57ff-4d58-8812-b72f624b7739\" (UID: \"f23a73ee-57ff-4d58-8812-b72f624b7739\") " Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.718156 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl" (OuterVolumeSpecName: "kube-api-access-wjdfl") pod "f23a73ee-57ff-4d58-8812-b72f624b7739" (UID: "f23a73ee-57ff-4d58-8812-b72f624b7739"). InnerVolumeSpecName "kube-api-access-wjdfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.736649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f23a73ee-57ff-4d58-8812-b72f624b7739" (UID: "f23a73ee-57ff-4d58-8812-b72f624b7739"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.761083 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data" (OuterVolumeSpecName: "config-data") pod "f23a73ee-57ff-4d58-8812-b72f624b7739" (UID: "f23a73ee-57ff-4d58-8812-b72f624b7739"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.782291 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:06 crc kubenswrapper[4907]: E0313 15:34:06.782907 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.815164 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjdfl\" (UniqueName: \"kubernetes.io/projected/f23a73ee-57ff-4d58-8812-b72f624b7739-kube-api-access-wjdfl\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.815205 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:06 crc kubenswrapper[4907]: I0313 15:34:06.815220 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f23a73ee-57ff-4d58-8812-b72f624b7739-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.191553 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4gtrw" event={"ID":"f23a73ee-57ff-4d58-8812-b72f624b7739","Type":"ContainerDied","Data":"3c42a8c243c0093bfc9f1f8cf6bfd418cc51eaed61e0ec45a9fbb1802050c2c9"} Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.191862 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c42a8c243c0093bfc9f1f8cf6bfd418cc51eaed61e0ec45a9fbb1802050c2c9" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.191618 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4gtrw" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427056 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:34:07 crc kubenswrapper[4907]: E0313 15:34:07.427683 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerName="oc" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427701 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerName="oc" Mar 13 15:34:07 crc kubenswrapper[4907]: E0313 15:34:07.427743 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerName="keystone-db-sync" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427753 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerName="keystone-db-sync" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.427960 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" containerName="keystone-db-sync" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.428000 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" containerName="oc" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.429166 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.442031 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.502940 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.504297 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.509801 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.510639 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.510826 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.511029 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.511187 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.511382 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.531899 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.531976 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.532107 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.532140 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.532166 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633677 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633731 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633783 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633806 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633824 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633862 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.633904 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.634020 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.634132 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.634817 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.637038 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.638367 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.638397 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.638420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.659755 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"dnsmasq-dns-6c5b6fdc47-txgc2\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.738921 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739304 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739340 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739372 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739393 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.739429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.742839 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.743067 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.743173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.745003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.745015 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.755324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.759721 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"keystone-bootstrap-mqwdv\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:07 crc kubenswrapper[4907]: I0313 15:34:07.834902 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:08 crc kubenswrapper[4907]: I0313 15:34:08.217913 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:34:08 crc kubenswrapper[4907]: W0313 15:34:08.221636 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc26e3cb6_4f00_49d9_8bc2_9ec347574580.slice/crio-5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5 WatchSource:0}: Error finding container 5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5: Status 404 returned error can't find the container with id 5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5 Mar 13 15:34:08 crc kubenswrapper[4907]: W0313 15:34:08.347160 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1838c955_2910_4589_a64f_7fe9f865d991.slice/crio-caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f WatchSource:0}: Error finding container caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f: Status 404 returned error can't find the container with id caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f Mar 13 15:34:08 crc kubenswrapper[4907]: I0313 15:34:08.351537 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.210275 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerStarted","Data":"47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.210967 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerStarted","Data":"caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.212198 4907 generic.go:334] "Generic (PLEG): container finished" podID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerID="de00b39199d31a11cec915e1367446617b6ca45fa3186e7f080356f4a1332db1" exitCode=0 Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.212236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerDied","Data":"de00b39199d31a11cec915e1367446617b6ca45fa3186e7f080356f4a1332db1"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.212260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerStarted","Data":"5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5"} Mar 13 15:34:09 crc kubenswrapper[4907]: I0313 15:34:09.236345 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mqwdv" podStartSLOduration=2.236323909 podStartE2EDuration="2.236323909s" podCreationTimestamp="2026-03-13 15:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:09.235769184 +0000 UTC m=+5348.135556873" watchObservedRunningTime="2026-03-13 15:34:09.236323909 +0000 UTC m=+5348.136111618" Mar 13 15:34:10 crc kubenswrapper[4907]: I0313 15:34:10.220030 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerStarted","Data":"f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2"} Mar 13 15:34:10 crc kubenswrapper[4907]: I0313 15:34:10.220654 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:10 crc kubenswrapper[4907]: I0313 15:34:10.248466 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" podStartSLOduration=3.248444712 podStartE2EDuration="3.248444712s" podCreationTimestamp="2026-03-13 15:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:10.24546877 +0000 UTC m=+5349.145256469" watchObservedRunningTime="2026-03-13 15:34:10.248444712 +0000 UTC m=+5349.148232411" Mar 13 15:34:11 crc kubenswrapper[4907]: I0313 15:34:11.475643 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 13 15:34:12 crc kubenswrapper[4907]: I0313 15:34:12.237138 4907 generic.go:334] "Generic (PLEG): container finished" podID="1838c955-2910-4589-a64f-7fe9f865d991" containerID="47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32" exitCode=0 Mar 13 15:34:12 crc kubenswrapper[4907]: I0313 15:34:12.237176 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerDied","Data":"47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32"} Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.624291 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785432 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785489 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785573 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785653 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785686 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.785977 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") pod \"1838c955-2910-4589-a64f-7fe9f865d991\" (UID: \"1838c955-2910-4589-a64f-7fe9f865d991\") " Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.791422 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.791446 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt" (OuterVolumeSpecName: "kube-api-access-kf2mt") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "kube-api-access-kf2mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.793712 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts" (OuterVolumeSpecName: "scripts") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.794470 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.811077 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.814492 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data" (OuterVolumeSpecName: "config-data") pod "1838c955-2910-4589-a64f-7fe9f865d991" (UID: "1838c955-2910-4589-a64f-7fe9f865d991"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888656 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888700 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888718 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888734 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888747 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf2mt\" (UniqueName: \"kubernetes.io/projected/1838c955-2910-4589-a64f-7fe9f865d991-kube-api-access-kf2mt\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:13 crc kubenswrapper[4907]: I0313 15:34:13.888762 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1838c955-2910-4589-a64f-7fe9f865d991-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.256174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwdv" event={"ID":"1838c955-2910-4589-a64f-7fe9f865d991","Type":"ContainerDied","Data":"caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f"} Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.256225 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caca426201cac3321e9d42ef962695a8f59b09bc5a76701b84aff5acd2e5bb8f" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.256272 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwdv" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.341090 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.350565 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mqwdv"] Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.416989 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:34:14 crc kubenswrapper[4907]: E0313 15:34:14.417360 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1838c955-2910-4589-a64f-7fe9f865d991" containerName="keystone-bootstrap" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.417384 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1838c955-2910-4589-a64f-7fe9f865d991" containerName="keystone-bootstrap" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.417589 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1838c955-2910-4589-a64f-7fe9f865d991" containerName="keystone-bootstrap" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.418256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.422332 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.422554 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.423915 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.423961 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.423970 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.428217 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600031 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600095 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600121 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600172 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600226 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.600261 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.700947 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701068 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701140 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.701189 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.705097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.705318 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.706210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.707069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.708245 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.717669 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"keystone-bootstrap-mf86w\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:14 crc kubenswrapper[4907]: I0313 15:34:14.734707 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:15 crc kubenswrapper[4907]: I0313 15:34:15.177544 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:34:15 crc kubenswrapper[4907]: I0313 15:34:15.267373 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerStarted","Data":"d609585b4e370b2fba0b235b1daafc7fcb8099494f72d1a80e4d09fbc6068dc3"} Mar 13 15:34:15 crc kubenswrapper[4907]: I0313 15:34:15.792419 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1838c955-2910-4589-a64f-7fe9f865d991" path="/var/lib/kubelet/pods/1838c955-2910-4589-a64f-7fe9f865d991/volumes" Mar 13 15:34:16 crc kubenswrapper[4907]: I0313 15:34:16.278237 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerStarted","Data":"e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905"} Mar 13 15:34:16 crc kubenswrapper[4907]: I0313 15:34:16.301625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mf86w" podStartSLOduration=2.301608373 podStartE2EDuration="2.301608373s" podCreationTimestamp="2026-03-13 15:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:16.29710399 +0000 UTC m=+5355.196891679" watchObservedRunningTime="2026-03-13 15:34:16.301608373 +0000 UTC m=+5355.201396062" Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.757091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.782904 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:17 crc kubenswrapper[4907]: E0313 15:34:17.783102 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.838852 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:34:17 crc kubenswrapper[4907]: I0313 15:34:17.839168 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" containerID="cri-o://a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965" gracePeriod=10 Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300129 4907 generic.go:334] "Generic (PLEG): container finished" podID="49e01f82-fb82-4351-a637-3b885e663e0e" containerID="a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965" exitCode=0 Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300230 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerDied","Data":"a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965"} Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300271 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" event={"ID":"49e01f82-fb82-4351-a637-3b885e663e0e","Type":"ContainerDied","Data":"2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101"} Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.300290 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cc3bc2c8a0a2c514a94d228c418e25f81f45826dc803a6c0203288a2ddf2101" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.302166 4907 generic.go:334] "Generic (PLEG): container finished" podID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerID="e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905" exitCode=0 Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.302199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerDied","Data":"e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905"} Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.329485 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463016 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463104 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463201 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.463252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") pod \"49e01f82-fb82-4351-a637-3b885e663e0e\" (UID: \"49e01f82-fb82-4351-a637-3b885e663e0e\") " Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.478154 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c" (OuterVolumeSpecName: "kube-api-access-lk79c") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "kube-api-access-lk79c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.504423 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.504474 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.516261 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.525079 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config" (OuterVolumeSpecName: "config") pod "49e01f82-fb82-4351-a637-3b885e663e0e" (UID: "49e01f82-fb82-4351-a637-3b885e663e0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565469 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565506 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565518 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565525 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e01f82-fb82-4351-a637-3b885e663e0e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:18 crc kubenswrapper[4907]: I0313 15:34:18.565535 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk79c\" (UniqueName: \"kubernetes.io/projected/49e01f82-fb82-4351-a637-3b885e663e0e-kube-api-access-lk79c\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.309925 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65b7b7b995-wl96j" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.349606 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.357395 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65b7b7b995-wl96j"] Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.627607 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.787984 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788024 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788091 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788124 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788177 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.788207 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") pod \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\" (UID: \"0056ce52-b92e-4f2d-b3b0-a7dff01207ff\") " Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.793478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.793970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts" (OuterVolumeSpecName: "scripts") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.794342 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf" (OuterVolumeSpecName: "kube-api-access-7qbqf") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "kube-api-access-7qbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.794398 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.799348 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" path="/var/lib/kubelet/pods/49e01f82-fb82-4351-a637-3b885e663e0e/volumes" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.810042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.811908 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data" (OuterVolumeSpecName: "config-data") pod "0056ce52-b92e-4f2d-b3b0-a7dff01207ff" (UID: "0056ce52-b92e-4f2d-b3b0-a7dff01207ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891233 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891681 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891710 4907 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891736 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qbqf\" (UniqueName: \"kubernetes.io/projected/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-kube-api-access-7qbqf\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891761 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:19 crc kubenswrapper[4907]: I0313 15:34:19.891783 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0056ce52-b92e-4f2d-b3b0-a7dff01207ff-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.321770 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mf86w" event={"ID":"0056ce52-b92e-4f2d-b3b0-a7dff01207ff","Type":"ContainerDied","Data":"d609585b4e370b2fba0b235b1daafc7fcb8099494f72d1a80e4d09fbc6068dc3"} Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.321820 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d609585b4e370b2fba0b235b1daafc7fcb8099494f72d1a80e4d09fbc6068dc3" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.322082 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mf86w" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.416850 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-58b5f55f68-nlrbn"] Mar 13 15:34:20 crc kubenswrapper[4907]: E0313 15:34:20.417958 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="init" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.417985 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="init" Mar 13 15:34:20 crc kubenswrapper[4907]: E0313 15:34:20.418026 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418034 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" Mar 13 15:34:20 crc kubenswrapper[4907]: E0313 15:34:20.418052 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerName="keystone-bootstrap" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418064 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerName="keystone-bootstrap" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418266 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" containerName="keystone-bootstrap" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418284 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e01f82-fb82-4351-a637-3b885e663e0e" containerName="dnsmasq-dns" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.418926 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.420989 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gjbw7" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.421002 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.421299 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.430378 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58b5f55f68-nlrbn"] Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.432249 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.601914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b987\" (UniqueName: \"kubernetes.io/projected/121ed53e-99ad-41a2-b8e6-56e4f56520c6-kube-api-access-8b987\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602176 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-fernet-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-credential-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602279 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-combined-ca-bundle\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602331 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-scripts\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.602397 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-config-data\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703504 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-config-data\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703592 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b987\" (UniqueName: \"kubernetes.io/projected/121ed53e-99ad-41a2-b8e6-56e4f56520c6-kube-api-access-8b987\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-fernet-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-credential-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.703757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-combined-ca-bundle\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.704302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-scripts\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.707777 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-config-data\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.714397 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-scripts\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.714396 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-credential-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.714701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-combined-ca-bundle\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.715411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/121ed53e-99ad-41a2-b8e6-56e4f56520c6-fernet-keys\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.729418 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b987\" (UniqueName: \"kubernetes.io/projected/121ed53e-99ad-41a2-b8e6-56e4f56520c6-kube-api-access-8b987\") pod \"keystone-58b5f55f68-nlrbn\" (UID: \"121ed53e-99ad-41a2-b8e6-56e4f56520c6\") " pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:20 crc kubenswrapper[4907]: I0313 15:34:20.738356 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:21 crc kubenswrapper[4907]: I0313 15:34:21.171018 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-58b5f55f68-nlrbn"] Mar 13 15:34:21 crc kubenswrapper[4907]: I0313 15:34:21.339107 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58b5f55f68-nlrbn" event={"ID":"121ed53e-99ad-41a2-b8e6-56e4f56520c6","Type":"ContainerStarted","Data":"47fb56f43f328a13f175803cd3cef1e5f392538a9ed6b02f2b0e2451c024895b"} Mar 13 15:34:22 crc kubenswrapper[4907]: I0313 15:34:22.350687 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-58b5f55f68-nlrbn" event={"ID":"121ed53e-99ad-41a2-b8e6-56e4f56520c6","Type":"ContainerStarted","Data":"62d31481988e6763e874bc01945c073a90df2def5706e0b04fd7660a70a1cf83"} Mar 13 15:34:22 crc kubenswrapper[4907]: I0313 15:34:22.350840 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:22 crc kubenswrapper[4907]: I0313 15:34:22.380764 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-58b5f55f68-nlrbn" podStartSLOduration=2.380739015 podStartE2EDuration="2.380739015s" podCreationTimestamp="2026-03-13 15:34:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:22.370347011 +0000 UTC m=+5361.270134700" watchObservedRunningTime="2026-03-13 15:34:22.380739015 +0000 UTC m=+5361.280526704" Mar 13 15:34:27 crc kubenswrapper[4907]: I0313 15:34:27.934717 4907 scope.go:117] "RemoveContainer" containerID="206bf50c83686d14437a447b0be4276ceccf75f1f0402647af114c3e1ce67140" Mar 13 15:34:29 crc kubenswrapper[4907]: I0313 15:34:29.781920 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:29 crc kubenswrapper[4907]: E0313 15:34:29.782478 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:44 crc kubenswrapper[4907]: I0313 15:34:44.782589 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:44 crc kubenswrapper[4907]: E0313 15:34:44.783338 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:34:52 crc kubenswrapper[4907]: I0313 15:34:52.168903 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-58b5f55f68-nlrbn" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.656203 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.657561 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.660032 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.660045 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.668009 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ff76w" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.680744 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.782707 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.859086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.859321 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.859438 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.963395 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.964100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.964166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.964934 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.969507 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:56 crc kubenswrapper[4907]: I0313 15:34:56.986545 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"openstackclient\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " pod="openstack/openstackclient" Mar 13 15:34:57 crc kubenswrapper[4907]: I0313 15:34:57.277437 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:34:57 crc kubenswrapper[4907]: I0313 15:34:57.649953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4"} Mar 13 15:34:57 crc kubenswrapper[4907]: I0313 15:34:57.719842 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:34:57 crc kubenswrapper[4907]: W0313 15:34:57.723027 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74a5c619_2dde_4f75_9106_cad6a39265ce.slice/crio-677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae WatchSource:0}: Error finding container 677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae: Status 404 returned error can't find the container with id 677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae Mar 13 15:34:58 crc kubenswrapper[4907]: I0313 15:34:58.658822 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"74a5c619-2dde-4f75-9106-cad6a39265ce","Type":"ContainerStarted","Data":"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff"} Mar 13 15:34:58 crc kubenswrapper[4907]: I0313 15:34:58.659129 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"74a5c619-2dde-4f75-9106-cad6a39265ce","Type":"ContainerStarted","Data":"677d1ff690530322a0aa44c9900ffa8ce8057cfe59c98e9b289689a68f5fd8ae"} Mar 13 15:34:58 crc kubenswrapper[4907]: I0313 15:34:58.679050 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.679003232 podStartE2EDuration="2.679003232s" podCreationTimestamp="2026-03-13 15:34:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:34:58.67455922 +0000 UTC m=+5397.574346949" watchObservedRunningTime="2026-03-13 15:34:58.679003232 +0000 UTC m=+5397.578790921" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.665758 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.668231 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.685780 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.799810 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.799863 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.799920 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.901645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902095 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902417 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902847 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.902845 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.920221 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"community-operators-vkhz4\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:50 crc kubenswrapper[4907]: I0313 15:35:50.987602 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:35:51 crc kubenswrapper[4907]: I0313 15:35:51.513795 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:35:52 crc kubenswrapper[4907]: I0313 15:35:52.100334 4907 generic.go:334] "Generic (PLEG): container finished" podID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" exitCode=0 Mar 13 15:35:52 crc kubenswrapper[4907]: I0313 15:35:52.100388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a"} Mar 13 15:35:52 crc kubenswrapper[4907]: I0313 15:35:52.100617 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerStarted","Data":"d2d3846d126727c38882c055157fb114b51e62e91302d46c000f07a01a3197ba"} Mar 13 15:35:53 crc kubenswrapper[4907]: I0313 15:35:53.114553 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerStarted","Data":"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5"} Mar 13 15:35:54 crc kubenswrapper[4907]: I0313 15:35:54.124819 4907 generic.go:334] "Generic (PLEG): container finished" podID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" exitCode=0 Mar 13 15:35:54 crc kubenswrapper[4907]: I0313 15:35:54.124962 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5"} Mar 13 15:35:55 crc kubenswrapper[4907]: I0313 15:35:55.133932 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerStarted","Data":"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8"} Mar 13 15:35:55 crc kubenswrapper[4907]: I0313 15:35:55.156010 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vkhz4" podStartSLOduration=2.527085146 podStartE2EDuration="5.155987781s" podCreationTimestamp="2026-03-13 15:35:50 +0000 UTC" firstStartedPulling="2026-03-13 15:35:52.102850594 +0000 UTC m=+5451.002638283" lastFinishedPulling="2026-03-13 15:35:54.731753229 +0000 UTC m=+5453.631540918" observedRunningTime="2026-03-13 15:35:55.149224606 +0000 UTC m=+5454.049012295" watchObservedRunningTime="2026-03-13 15:35:55.155987781 +0000 UTC m=+5454.055775460" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.147552 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.150546 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.153683 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.155099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.158583 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.172947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.252251 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"auto-csr-approver-29556936-bwsf9\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.354763 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"auto-csr-approver-29556936-bwsf9\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.398501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"auto-csr-approver-29556936-bwsf9\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.477572 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.903057 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.988530 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:00 crc kubenswrapper[4907]: I0313 15:36:00.988570 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:01 crc kubenswrapper[4907]: I0313 15:36:01.028008 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:01 crc kubenswrapper[4907]: I0313 15:36:01.191048 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" event={"ID":"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7","Type":"ContainerStarted","Data":"8ffd45b87adb147d65f6690cb310171e5230c18b1399582fbe6807ed0ee79621"} Mar 13 15:36:01 crc kubenswrapper[4907]: I0313 15:36:01.228700 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:02 crc kubenswrapper[4907]: I0313 15:36:02.164948 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.216672 4907 generic.go:334] "Generic (PLEG): container finished" podID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerID="a02ae219ab9b5a1fd678d76ac3b1c7e41a88bec1b6ae3e7a16334b35c97e0864" exitCode=0 Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.216789 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" event={"ID":"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7","Type":"ContainerDied","Data":"a02ae219ab9b5a1fd678d76ac3b1c7e41a88bec1b6ae3e7a16334b35c97e0864"} Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.217617 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vkhz4" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" containerID="cri-o://b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" gracePeriod=2 Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.649855 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.709992 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") pod \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.710138 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") pod \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.710313 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") pod \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\" (UID: \"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa\") " Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.711166 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities" (OuterVolumeSpecName: "utilities") pod "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" (UID: "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.717341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m" (OuterVolumeSpecName: "kube-api-access-w7d4m") pod "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" (UID: "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa"). InnerVolumeSpecName "kube-api-access-w7d4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.769740 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" (UID: "7213ba2d-c84c-43f8-99a6-c08f87c5c0aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.812799 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7d4m\" (UniqueName: \"kubernetes.io/projected/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-kube-api-access-w7d4m\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.812825 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:03 crc kubenswrapper[4907]: I0313 15:36:03.812834 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.225957 4907 generic.go:334] "Generic (PLEG): container finished" podID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" exitCode=0 Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226019 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vkhz4" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8"} Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vkhz4" event={"ID":"7213ba2d-c84c-43f8-99a6-c08f87c5c0aa","Type":"ContainerDied","Data":"d2d3846d126727c38882c055157fb114b51e62e91302d46c000f07a01a3197ba"} Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.226135 4907 scope.go:117] "RemoveContainer" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.252988 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.258126 4907 scope.go:117] "RemoveContainer" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.261456 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vkhz4"] Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.306937 4907 scope.go:117] "RemoveContainer" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.323472 4907 scope.go:117] "RemoveContainer" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" Mar 13 15:36:04 crc kubenswrapper[4907]: E0313 15:36:04.324509 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8\": container with ID starting with b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8 not found: ID does not exist" containerID="b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.324557 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8"} err="failed to get container status \"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8\": rpc error: code = NotFound desc = could not find container \"b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8\": container with ID starting with b0d053e1bf868544e4714cfc718cdfdb940c35699b616905f2247bf747bbb8a8 not found: ID does not exist" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.324590 4907 scope.go:117] "RemoveContainer" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" Mar 13 15:36:04 crc kubenswrapper[4907]: E0313 15:36:04.324997 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5\": container with ID starting with 57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5 not found: ID does not exist" containerID="57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.325097 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5"} err="failed to get container status \"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5\": rpc error: code = NotFound desc = could not find container \"57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5\": container with ID starting with 57388821445c4b768505f257c2ec6fe32c44ef455b2d02ef9106d2a78dd49da5 not found: ID does not exist" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.325142 4907 scope.go:117] "RemoveContainer" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" Mar 13 15:36:04 crc kubenswrapper[4907]: E0313 15:36:04.325837 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a\": container with ID starting with 378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a not found: ID does not exist" containerID="378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.325869 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a"} err="failed to get container status \"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a\": rpc error: code = NotFound desc = could not find container \"378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a\": container with ID starting with 378658932bb4f93263f9c7c834e5eb6a55511ab6e84b9fd5f0fe1ea37d58141a not found: ID does not exist" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.572595 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.628788 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") pod \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\" (UID: \"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7\") " Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.635110 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh" (OuterVolumeSpecName: "kube-api-access-ngwfh") pod "38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" (UID: "38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7"). InnerVolumeSpecName "kube-api-access-ngwfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:04 crc kubenswrapper[4907]: I0313 15:36:04.730755 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngwfh\" (UniqueName: \"kubernetes.io/projected/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7-kube-api-access-ngwfh\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.235174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" event={"ID":"38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7","Type":"ContainerDied","Data":"8ffd45b87adb147d65f6690cb310171e5230c18b1399582fbe6807ed0ee79621"} Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.235215 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ffd45b87adb147d65f6690cb310171e5230c18b1399582fbe6807ed0ee79621" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.235231 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556936-bwsf9" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.662568 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.670214 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556930-ndtqs"] Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.791797 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" path="/var/lib/kubelet/pods/7213ba2d-c84c-43f8-99a6-c08f87c5c0aa/volumes" Mar 13 15:36:05 crc kubenswrapper[4907]: I0313 15:36:05.792589 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df83f491-eb7d-4e1c-b814-28cc7ef9abee" path="/var/lib/kubelet/pods/df83f491-eb7d-4e1c-b814-28cc7ef9abee/volumes" Mar 13 15:36:28 crc kubenswrapper[4907]: I0313 15:36:28.062300 4907 scope.go:117] "RemoveContainer" containerID="9c966e9f83d2d89ea300010e1e5e03c7e34fbdc9099449a4ffcb481299d87028" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.463689 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464735 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-content" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464752 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-content" Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464793 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-utilities" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464802 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="extract-utilities" Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464833 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerName="oc" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464842 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerName="oc" Mar 13 15:36:40 crc kubenswrapper[4907]: E0313 15:36:40.464858 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.464866 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.465110 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7213ba2d-c84c-43f8-99a6-c08f87c5c0aa" containerName="registry-server" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.465136 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" containerName="oc" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.465938 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.471417 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.473176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.475892 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.499508 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.526713 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670083 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670168 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.670313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.771664 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.771984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772783 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.772871 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.789208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"barbican-b654-account-create-update-4jjkj\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.795545 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"barbican-db-create-4jnct\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.795974 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:40 crc kubenswrapper[4907]: I0313 15:36:40.812145 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.242901 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.323574 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:36:41 crc kubenswrapper[4907]: W0313 15:36:41.334486 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod523d3b7c_7098_44bc_92a6_d48f3ae49b65.slice/crio-56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a WatchSource:0}: Error finding container 56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a: Status 404 returned error can't find the container with id 56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.545623 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerStarted","Data":"0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.545665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerStarted","Data":"fd2873409a27ed99fa509e75a3b1acb251ee1ed6a6086a5abc56230de487ac8f"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.548548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerStarted","Data":"ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.548599 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerStarted","Data":"56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a"} Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.560013 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-b654-account-create-update-4jjkj" podStartSLOduration=1.559994263 podStartE2EDuration="1.559994263s" podCreationTimestamp="2026-03-13 15:36:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:41.558753269 +0000 UTC m=+5500.458540978" watchObservedRunningTime="2026-03-13 15:36:41.559994263 +0000 UTC m=+5500.459781952" Mar 13 15:36:41 crc kubenswrapper[4907]: I0313 15:36:41.572601 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-4jnct" podStartSLOduration=1.572582347 podStartE2EDuration="1.572582347s" podCreationTimestamp="2026-03-13 15:36:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:41.572348831 +0000 UTC m=+5500.472136520" watchObservedRunningTime="2026-03-13 15:36:41.572582347 +0000 UTC m=+5500.472370036" Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.559348 4907 generic.go:334] "Generic (PLEG): container finished" podID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerID="0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a" exitCode=0 Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.559443 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerDied","Data":"0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a"} Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.561636 4907 generic.go:334] "Generic (PLEG): container finished" podID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerID="ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6" exitCode=0 Mar 13 15:36:42 crc kubenswrapper[4907]: I0313 15:36:42.561697 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerDied","Data":"ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6"} Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.920119 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.927391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") pod \"1e73852b-fdd0-4cd2-a268-96f6ed551011\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.927498 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") pod \"1e73852b-fdd0-4cd2-a268-96f6ed551011\" (UID: \"1e73852b-fdd0-4cd2-a268-96f6ed551011\") " Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.929539 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e73852b-fdd0-4cd2-a268-96f6ed551011" (UID: "1e73852b-fdd0-4cd2-a268-96f6ed551011"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.933027 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2" (OuterVolumeSpecName: "kube-api-access-rwpg2") pod "1e73852b-fdd0-4cd2-a268-96f6ed551011" (UID: "1e73852b-fdd0-4cd2-a268-96f6ed551011"). InnerVolumeSpecName "kube-api-access-rwpg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:43 crc kubenswrapper[4907]: I0313 15:36:43.935312 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.029293 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwpg2\" (UniqueName: \"kubernetes.io/projected/1e73852b-fdd0-4cd2-a268-96f6ed551011-kube-api-access-rwpg2\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.029333 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e73852b-fdd0-4cd2-a268-96f6ed551011-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.130194 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") pod \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.130285 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") pod \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\" (UID: \"523d3b7c-7098-44bc-92a6-d48f3ae49b65\") " Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.130986 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "523d3b7c-7098-44bc-92a6-d48f3ae49b65" (UID: "523d3b7c-7098-44bc-92a6-d48f3ae49b65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.135420 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8" (OuterVolumeSpecName: "kube-api-access-zkhl8") pod "523d3b7c-7098-44bc-92a6-d48f3ae49b65" (UID: "523d3b7c-7098-44bc-92a6-d48f3ae49b65"). InnerVolumeSpecName "kube-api-access-zkhl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.232654 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkhl8\" (UniqueName: \"kubernetes.io/projected/523d3b7c-7098-44bc-92a6-d48f3ae49b65-kube-api-access-zkhl8\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.232719 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/523d3b7c-7098-44bc-92a6-d48f3ae49b65-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.576947 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b654-account-create-update-4jjkj" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.576943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b654-account-create-update-4jjkj" event={"ID":"1e73852b-fdd0-4cd2-a268-96f6ed551011","Type":"ContainerDied","Data":"fd2873409a27ed99fa509e75a3b1acb251ee1ed6a6086a5abc56230de487ac8f"} Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.577084 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd2873409a27ed99fa509e75a3b1acb251ee1ed6a6086a5abc56230de487ac8f" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.578944 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4jnct" event={"ID":"523d3b7c-7098-44bc-92a6-d48f3ae49b65","Type":"ContainerDied","Data":"56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a"} Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.578969 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4jnct" Mar 13 15:36:44 crc kubenswrapper[4907]: I0313 15:36:44.578990 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56a67057374a475d911f887833cd0ea7420af1041fbdd3d80eddab806d07786a" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.677715 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:36:45 crc kubenswrapper[4907]: E0313 15:36:45.678306 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerName="mariadb-database-create" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678321 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerName="mariadb-database-create" Mar 13 15:36:45 crc kubenswrapper[4907]: E0313 15:36:45.678350 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerName="mariadb-account-create-update" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678358 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerName="mariadb-account-create-update" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678575 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" containerName="mariadb-account-create-update" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.678597 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" containerName="mariadb-database-create" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.679294 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.681384 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9gtqn" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.686658 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.690277 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.757943 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.758004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.758117 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.859087 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.859204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.859265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.864780 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.864799 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.893562 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"barbican-db-sync-xrjlt\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:45 crc kubenswrapper[4907]: I0313 15:36:45.995627 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:46 crc kubenswrapper[4907]: I0313 15:36:46.614636 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:36:47 crc kubenswrapper[4907]: I0313 15:36:47.602053 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerStarted","Data":"29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef"} Mar 13 15:36:47 crc kubenswrapper[4907]: I0313 15:36:47.603473 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerStarted","Data":"d8afe419c9f005bbe4627b889eb2b0b424eb7134bbdb9b7658ade7f6e310e864"} Mar 13 15:36:47 crc kubenswrapper[4907]: I0313 15:36:47.624560 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xrjlt" podStartSLOduration=2.624541947 podStartE2EDuration="2.624541947s" podCreationTimestamp="2026-03-13 15:36:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:47.618540452 +0000 UTC m=+5506.518328161" watchObservedRunningTime="2026-03-13 15:36:47.624541947 +0000 UTC m=+5506.524329656" Mar 13 15:36:48 crc kubenswrapper[4907]: I0313 15:36:48.611507 4907 generic.go:334] "Generic (PLEG): container finished" podID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerID="29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef" exitCode=0 Mar 13 15:36:48 crc kubenswrapper[4907]: I0313 15:36:48.611548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerDied","Data":"29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef"} Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.083192 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.089614 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-l8vbt"] Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.791228 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d5a861d-b4ca-4936-9fac-5aa23fcaa317" path="/var/lib/kubelet/pods/1d5a861d-b4ca-4936-9fac-5aa23fcaa317/volumes" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.915116 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.946862 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") pod \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.947133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") pod \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.947179 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") pod \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\" (UID: \"df0d68f5-c073-4dff-8f70-f2897aaef9dc\") " Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.952595 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "df0d68f5-c073-4dff-8f70-f2897aaef9dc" (UID: "df0d68f5-c073-4dff-8f70-f2897aaef9dc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.953465 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q" (OuterVolumeSpecName: "kube-api-access-hdl8q") pod "df0d68f5-c073-4dff-8f70-f2897aaef9dc" (UID: "df0d68f5-c073-4dff-8f70-f2897aaef9dc"). InnerVolumeSpecName "kube-api-access-hdl8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:36:49 crc kubenswrapper[4907]: I0313 15:36:49.973346 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df0d68f5-c073-4dff-8f70-f2897aaef9dc" (UID: "df0d68f5-c073-4dff-8f70-f2897aaef9dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.049234 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.049270 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df0d68f5-c073-4dff-8f70-f2897aaef9dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.049280 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdl8q\" (UniqueName: \"kubernetes.io/projected/df0d68f5-c073-4dff-8f70-f2897aaef9dc-kube-api-access-hdl8q\") on node \"crc\" DevicePath \"\"" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.632140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xrjlt" event={"ID":"df0d68f5-c073-4dff-8f70-f2897aaef9dc","Type":"ContainerDied","Data":"d8afe419c9f005bbe4627b889eb2b0b424eb7134bbdb9b7658ade7f6e310e864"} Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.632488 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8afe419c9f005bbe4627b889eb2b0b424eb7134bbdb9b7658ade7f6e310e864" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.632187 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xrjlt" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.854352 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-68d9445c58-bjlh2"] Mar 13 15:36:50 crc kubenswrapper[4907]: E0313 15:36:50.856039 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerName="barbican-db-sync" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.857157 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerName="barbican-db-sync" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.857574 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" containerName="barbican-db-sync" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.858705 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.861409 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-868b698d7f-wbtl4"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.862438 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-9gtqn" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.862870 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.862942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.863032 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.865080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.880153 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-868b698d7f-wbtl4"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.895116 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68d9445c58-bjlh2"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.963947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zts7\" (UniqueName: \"kubernetes.io/projected/3599175f-0bbd-4645-ab11-07e26d7fb98d-kube-api-access-6zts7\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964088 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b4d8b9-57da-4c61-9dd9-d746fc62db84-logs\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964111 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964142 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3599175f-0bbd-4645-ab11-07e26d7fb98d-logs\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964161 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-combined-ca-bundle\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964177 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rqph\" (UniqueName: \"kubernetes.io/projected/87b4d8b9-57da-4c61-9dd9-d746fc62db84-kube-api-access-9rqph\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964215 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data-custom\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964230 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964244 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data-custom\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.964267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-combined-ca-bundle\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.980977 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.982585 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:50 crc kubenswrapper[4907]: I0313 15:36:50.989071 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.060801 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5564857dbd-vn7xm"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.062109 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068563 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data-custom\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068611 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068639 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data-custom\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068701 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068729 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-combined-ca-bundle\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068773 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068823 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zts7\" (UniqueName: \"kubernetes.io/projected/3599175f-0bbd-4645-ab11-07e26d7fb98d-kube-api-access-6zts7\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068920 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068951 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b4d8b9-57da-4c61-9dd9-d746fc62db84-logs\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.068976 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069038 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3599175f-0bbd-4645-ab11-07e26d7fb98d-logs\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069059 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-combined-ca-bundle\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069082 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rqph\" (UniqueName: \"kubernetes.io/projected/87b4d8b9-57da-4c61-9dd9-d746fc62db84-kube-api-access-9rqph\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.069106 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.070731 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b4d8b9-57da-4c61-9dd9-d746fc62db84-logs\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.077029 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3599175f-0bbd-4645-ab11-07e26d7fb98d-logs\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.077410 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.084345 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5564857dbd-vn7xm"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.088194 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data-custom\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.088584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-config-data\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.088679 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b4d8b9-57da-4c61-9dd9-d746fc62db84-combined-ca-bundle\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.091091 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data-custom\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.093578 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-combined-ca-bundle\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.094345 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3599175f-0bbd-4645-ab11-07e26d7fb98d-config-data\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.096563 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rqph\" (UniqueName: \"kubernetes.io/projected/87b4d8b9-57da-4c61-9dd9-d746fc62db84-kube-api-access-9rqph\") pod \"barbican-keystone-listener-68d9445c58-bjlh2\" (UID: \"87b4d8b9-57da-4c61-9dd9-d746fc62db84\") " pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.098391 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zts7\" (UniqueName: \"kubernetes.io/projected/3599175f-0bbd-4645-ab11-07e26d7fb98d-kube-api-access-6zts7\") pod \"barbican-worker-868b698d7f-wbtl4\" (UID: \"3599175f-0bbd-4645-ab11-07e26d7fb98d\") " pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170678 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data-custom\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170720 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170815 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170840 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170872 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170905 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-logs\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170927 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzlbb\" (UniqueName: \"kubernetes.io/projected/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-kube-api-access-jzlbb\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.170972 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-combined-ca-bundle\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.171782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.172602 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.172782 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.172862 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.176592 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.185728 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-868b698d7f-wbtl4" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.190356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"dnsmasq-dns-bd6bb59b7-mfqq8\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-logs\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzlbb\" (UniqueName: \"kubernetes.io/projected/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-kube-api-access-jzlbb\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272526 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-combined-ca-bundle\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data-custom\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.272627 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.273192 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-logs\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.279749 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.280305 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-config-data-custom\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.280306 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-combined-ca-bundle\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.296725 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.298247 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzlbb\" (UniqueName: \"kubernetes.io/projected/d83fca2b-ed84-4516-8510-6e0ff4b2cfef-kube-api-access-jzlbb\") pod \"barbican-api-5564857dbd-vn7xm\" (UID: \"d83fca2b-ed84-4516-8510-6e0ff4b2cfef\") " pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.460571 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.696349 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68d9445c58-bjlh2"] Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.706788 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-868b698d7f-wbtl4"] Mar 13 15:36:51 crc kubenswrapper[4907]: W0313 15:36:51.724901 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3599175f_0bbd_4645_ab11_07e26d7fb98d.slice/crio-70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717 WatchSource:0}: Error finding container 70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717: Status 404 returned error can't find the container with id 70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717 Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.862612 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:36:51 crc kubenswrapper[4907]: W0313 15:36:51.867480 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96134178_5784_41c4_bfb2_da280292c4ee.slice/crio-feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0 WatchSource:0}: Error finding container feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0: Status 404 returned error can't find the container with id feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0 Mar 13 15:36:51 crc kubenswrapper[4907]: I0313 15:36:51.971366 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5564857dbd-vn7xm"] Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650279 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5564857dbd-vn7xm" event={"ID":"d83fca2b-ed84-4516-8510-6e0ff4b2cfef","Type":"ContainerStarted","Data":"eb9beef76927f7ed790dbd2612952eb33c40a8f5966b2bc7ca005b02f2176cd0"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650769 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650845 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5564857dbd-vn7xm" event={"ID":"d83fca2b-ed84-4516-8510-6e0ff4b2cfef","Type":"ContainerStarted","Data":"a05c32904e57c5a2b807e53693b620924bb5ac41933bd588aad9c8e26572947b"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.650937 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5564857dbd-vn7xm" event={"ID":"d83fca2b-ed84-4516-8510-6e0ff4b2cfef","Type":"ContainerStarted","Data":"9e14830182c886273f0d1485e37dea5076045dd81323eaacf2d8eec0acdc38e9"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.652449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" event={"ID":"87b4d8b9-57da-4c61-9dd9-d746fc62db84","Type":"ContainerStarted","Data":"a5db3d582632269b451134f4ed459f116cf65a1a60b5091612a4c60f98350fb9"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.652491 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" event={"ID":"87b4d8b9-57da-4c61-9dd9-d746fc62db84","Type":"ContainerStarted","Data":"185cd70b5b58c41af371e1d5271c3806e31be590b16086db7b6711d2e98cc6ce"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.652502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" event={"ID":"87b4d8b9-57da-4c61-9dd9-d746fc62db84","Type":"ContainerStarted","Data":"e0c2059048ef60d27cd987200d115afaea617344cf5c9390a36015f744db9317"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.654704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-868b698d7f-wbtl4" event={"ID":"3599175f-0bbd-4645-ab11-07e26d7fb98d","Type":"ContainerStarted","Data":"0264c77b446d8dbb95b35312638da246b5431cf6b293a9faa6a0099662264a46"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.654742 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-868b698d7f-wbtl4" event={"ID":"3599175f-0bbd-4645-ab11-07e26d7fb98d","Type":"ContainerStarted","Data":"7c0f7e2f1cdaaf04ee701eb036ae9ee0a5c60418c67eb89e7d28d3e4b3d2e5ae"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.654759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-868b698d7f-wbtl4" event={"ID":"3599175f-0bbd-4645-ab11-07e26d7fb98d","Type":"ContainerStarted","Data":"70f8f900b1131883a154adee5909b47faa634913da47d0ca1dcba98372e75717"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.656339 4907 generic.go:334] "Generic (PLEG): container finished" podID="96134178-5784-41c4-bfb2-da280292c4ee" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" exitCode=0 Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.656379 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerDied","Data":"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.656399 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerStarted","Data":"feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0"} Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.682867 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5564857dbd-vn7xm" podStartSLOduration=1.682836918 podStartE2EDuration="1.682836918s" podCreationTimestamp="2026-03-13 15:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:52.67739928 +0000 UTC m=+5511.577186969" watchObservedRunningTime="2026-03-13 15:36:52.682836918 +0000 UTC m=+5511.582624607" Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.727332 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-868b698d7f-wbtl4" podStartSLOduration=2.727311493 podStartE2EDuration="2.727311493s" podCreationTimestamp="2026-03-13 15:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:52.723436467 +0000 UTC m=+5511.623224166" watchObservedRunningTime="2026-03-13 15:36:52.727311493 +0000 UTC m=+5511.627099182" Mar 13 15:36:52 crc kubenswrapper[4907]: I0313 15:36:52.759353 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-68d9445c58-bjlh2" podStartSLOduration=2.759335868 podStartE2EDuration="2.759335868s" podCreationTimestamp="2026-03-13 15:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:52.758960898 +0000 UTC m=+5511.658748587" watchObservedRunningTime="2026-03-13 15:36:52.759335868 +0000 UTC m=+5511.659123557" Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.666474 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerStarted","Data":"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2"} Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.667860 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.667993 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:36:53 crc kubenswrapper[4907]: I0313 15:36:53.686540 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" podStartSLOduration=3.686390327 podStartE2EDuration="3.686390327s" podCreationTimestamp="2026-03-13 15:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:36:53.681791681 +0000 UTC m=+5512.581579380" watchObservedRunningTime="2026-03-13 15:36:53.686390327 +0000 UTC m=+5512.586178026" Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.299131 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.361632 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.361929 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" containerID="cri-o://f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2" gracePeriod=10 Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.740086 4907 generic.go:334] "Generic (PLEG): container finished" podID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerID="f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2" exitCode=0 Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.740175 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerDied","Data":"f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2"} Mar 13 15:37:01 crc kubenswrapper[4907]: I0313 15:37:01.940111 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.062472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063136 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063376 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.063409 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") pod \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\" (UID: \"c26e3cb6-4f00-49d9-8bc2-9ec347574580\") " Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.073100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv" (OuterVolumeSpecName: "kube-api-access-hmslv") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "kube-api-access-hmslv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.108019 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.108932 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.122304 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config" (OuterVolumeSpecName: "config") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.132539 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c26e3cb6-4f00-49d9-8bc2-9ec347574580" (UID: "c26e3cb6-4f00-49d9-8bc2-9ec347574580"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166015 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166066 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmslv\" (UniqueName: \"kubernetes.io/projected/c26e3cb6-4f00-49d9-8bc2-9ec347574580-kube-api-access-hmslv\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166081 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166098 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.166109 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c26e3cb6-4f00-49d9-8bc2-9ec347574580-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.750365 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" event={"ID":"c26e3cb6-4f00-49d9-8bc2-9ec347574580","Type":"ContainerDied","Data":"5651c9bf703897955b69287b03043d49d9012737130119f792c71677cc9d3fe5"} Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.750429 4907 scope.go:117] "RemoveContainer" containerID="f3459463f090073ba747ae28c0f5e112ee217c5b14b17182f95119ff18ca2ea2" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.750586 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c5b6fdc47-txgc2" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.773690 4907 scope.go:117] "RemoveContainer" containerID="de00b39199d31a11cec915e1367446617b6ca45fa3186e7f080356f4a1332db1" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.789683 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.807682 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c5b6fdc47-txgc2"] Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.979867 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:37:02 crc kubenswrapper[4907]: I0313 15:37:02.998795 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5564857dbd-vn7xm" Mar 13 15:37:03 crc kubenswrapper[4907]: I0313 15:37:03.816294 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" path="/var/lib/kubelet/pods/c26e3cb6-4f00-49d9-8bc2-9ec347574580/volumes" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.633811 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:37:14 crc kubenswrapper[4907]: E0313 15:37:14.634587 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="init" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.634599 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="init" Mar 13 15:37:14 crc kubenswrapper[4907]: E0313 15:37:14.634610 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.634616 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.634767 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c26e3cb6-4f00-49d9-8bc2-9ec347574580" containerName="dnsmasq-dns" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.635334 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.647752 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.704600 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.704704 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.738952 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.740134 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.742306 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.752334 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806146 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806352 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.806970 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.830526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"neutron-db-create-jh7h4\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.907845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.908201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.908794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.940933 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"neutron-629b-account-create-update-rptv2\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:14 crc kubenswrapper[4907]: I0313 15:37:14.962276 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.055069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.461781 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.563641 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:37:15 crc kubenswrapper[4907]: W0313 15:37:15.569019 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod608c9419_22d5_4915_b5c6_8be70eebdb23.slice/crio-72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a WatchSource:0}: Error finding container 72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a: Status 404 returned error can't find the container with id 72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.872264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerStarted","Data":"2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.872650 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerStarted","Data":"72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.876028 4907 generic.go:334] "Generic (PLEG): container finished" podID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerID="6b2aa6e8ccbdaa4de43a0268ebdb8841de04a3498e495de6ed723c183300253b" exitCode=0 Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.876070 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7h4" event={"ID":"3cbbccb7-b0c0-49be-a969-ffa18b07bd13","Type":"ContainerDied","Data":"6b2aa6e8ccbdaa4de43a0268ebdb8841de04a3498e495de6ed723c183300253b"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.876087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7h4" event={"ID":"3cbbccb7-b0c0-49be-a969-ffa18b07bd13","Type":"ContainerStarted","Data":"874bae6b236dfc82c48bb081ac8b4f317afe5f9203d0173666c9a64b2661229b"} Mar 13 15:37:15 crc kubenswrapper[4907]: I0313 15:37:15.890133 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-629b-account-create-update-rptv2" podStartSLOduration=1.890108781 podStartE2EDuration="1.890108781s" podCreationTimestamp="2026-03-13 15:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:15.888356863 +0000 UTC m=+5534.788144562" watchObservedRunningTime="2026-03-13 15:37:15.890108781 +0000 UTC m=+5534.789896470" Mar 13 15:37:16 crc kubenswrapper[4907]: I0313 15:37:16.887961 4907 generic.go:334] "Generic (PLEG): container finished" podID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerID="2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9" exitCode=0 Mar 13 15:37:16 crc kubenswrapper[4907]: I0313 15:37:16.888161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerDied","Data":"2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9"} Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.283187 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.347851 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") pod \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.348101 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") pod \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\" (UID: \"3cbbccb7-b0c0-49be-a969-ffa18b07bd13\") " Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.348749 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3cbbccb7-b0c0-49be-a969-ffa18b07bd13" (UID: "3cbbccb7-b0c0-49be-a969-ffa18b07bd13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.353385 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b" (OuterVolumeSpecName: "kube-api-access-7hv4b") pod "3cbbccb7-b0c0-49be-a969-ffa18b07bd13" (UID: "3cbbccb7-b0c0-49be-a969-ffa18b07bd13"). InnerVolumeSpecName "kube-api-access-7hv4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.449751 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hv4b\" (UniqueName: \"kubernetes.io/projected/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-kube-api-access-7hv4b\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.449792 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3cbbccb7-b0c0-49be-a969-ffa18b07bd13-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.897387 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-jh7h4" Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.898049 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-jh7h4" event={"ID":"3cbbccb7-b0c0-49be-a969-ffa18b07bd13","Type":"ContainerDied","Data":"874bae6b236dfc82c48bb081ac8b4f317afe5f9203d0173666c9a64b2661229b"} Mar 13 15:37:17 crc kubenswrapper[4907]: I0313 15:37:17.898124 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="874bae6b236dfc82c48bb081ac8b4f317afe5f9203d0173666c9a64b2661229b" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.043993 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.044298 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.200691 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.265605 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") pod \"608c9419-22d5-4915-b5c6-8be70eebdb23\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.265780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") pod \"608c9419-22d5-4915-b5c6-8be70eebdb23\" (UID: \"608c9419-22d5-4915-b5c6-8be70eebdb23\") " Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.266183 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "608c9419-22d5-4915-b5c6-8be70eebdb23" (UID: "608c9419-22d5-4915-b5c6-8be70eebdb23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.270572 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd" (OuterVolumeSpecName: "kube-api-access-qh4pd") pod "608c9419-22d5-4915-b5c6-8be70eebdb23" (UID: "608c9419-22d5-4915-b5c6-8be70eebdb23"). InnerVolumeSpecName "kube-api-access-qh4pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.367228 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/608c9419-22d5-4915-b5c6-8be70eebdb23-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.367262 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh4pd\" (UniqueName: \"kubernetes.io/projected/608c9419-22d5-4915-b5c6-8be70eebdb23-kube-api-access-qh4pd\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.921389 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-629b-account-create-update-rptv2" event={"ID":"608c9419-22d5-4915-b5c6-8be70eebdb23","Type":"ContainerDied","Data":"72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a"} Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.921798 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72837ec872dcb1aeb63d3b6252b15b3fff4ae6b7c18ef8ec42f445d9fd0a0e5a" Mar 13 15:37:18 crc kubenswrapper[4907]: I0313 15:37:18.922148 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-629b-account-create-update-rptv2" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.955841 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:37:19 crc kubenswrapper[4907]: E0313 15:37:19.956251 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerName="mariadb-account-create-update" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956267 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerName="mariadb-account-create-update" Mar 13 15:37:19 crc kubenswrapper[4907]: E0313 15:37:19.956277 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerName="mariadb-database-create" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956284 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerName="mariadb-database-create" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956493 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" containerName="mariadb-account-create-update" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.956517 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" containerName="mariadb-database-create" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.957158 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.966716 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.966940 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rdwdw" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.967178 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 15:37:19 crc kubenswrapper[4907]: I0313 15:37:19.968252 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.005748 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.006353 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.006552 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.108535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.108599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.108704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.112343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.112444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.123569 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"neutron-db-sync-jtsxr\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.278909 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.751032 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:37:20 crc kubenswrapper[4907]: I0313 15:37:20.938875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerStarted","Data":"970fccf6e2bfb3136160282e85ca44140a4f0863ea34492c0b1a3eb07ca112e9"} Mar 13 15:37:21 crc kubenswrapper[4907]: I0313 15:37:21.948197 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerStarted","Data":"eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31"} Mar 13 15:37:21 crc kubenswrapper[4907]: I0313 15:37:21.962924 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jtsxr" podStartSLOduration=2.96286247 podStartE2EDuration="2.96286247s" podCreationTimestamp="2026-03-13 15:37:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:21.961194463 +0000 UTC m=+5540.860982152" watchObservedRunningTime="2026-03-13 15:37:21.96286247 +0000 UTC m=+5540.862650149" Mar 13 15:37:24 crc kubenswrapper[4907]: I0313 15:37:24.991357 4907 generic.go:334] "Generic (PLEG): container finished" podID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerID="eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31" exitCode=0 Mar 13 15:37:24 crc kubenswrapper[4907]: I0313 15:37:24.991437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerDied","Data":"eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31"} Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.292442 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.340704 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") pod \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.340816 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") pod \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.340860 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") pod \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\" (UID: \"1af3ec26-0a7c-44c4-8073-fc93093c0cf7\") " Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.346541 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5" (OuterVolumeSpecName: "kube-api-access-4vzb5") pod "1af3ec26-0a7c-44c4-8073-fc93093c0cf7" (UID: "1af3ec26-0a7c-44c4-8073-fc93093c0cf7"). InnerVolumeSpecName "kube-api-access-4vzb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.364275 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1af3ec26-0a7c-44c4-8073-fc93093c0cf7" (UID: "1af3ec26-0a7c-44c4-8073-fc93093c0cf7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.366995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config" (OuterVolumeSpecName: "config") pod "1af3ec26-0a7c-44c4-8073-fc93093c0cf7" (UID: "1af3ec26-0a7c-44c4-8073-fc93093c0cf7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.443013 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vzb5\" (UniqueName: \"kubernetes.io/projected/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-kube-api-access-4vzb5\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.443063 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:26 crc kubenswrapper[4907]: I0313 15:37:26.443075 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1af3ec26-0a7c-44c4-8073-fc93093c0cf7-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.011334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtsxr" event={"ID":"1af3ec26-0a7c-44c4-8073-fc93093c0cf7","Type":"ContainerDied","Data":"970fccf6e2bfb3136160282e85ca44140a4f0863ea34492c0b1a3eb07ca112e9"} Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.011372 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="970fccf6e2bfb3136160282e85ca44140a4f0863ea34492c0b1a3eb07ca112e9" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.011427 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtsxr" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.239212 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:37:27 crc kubenswrapper[4907]: E0313 15:37:27.240157 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerName="neutron-db-sync" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.240181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerName="neutron-db-sync" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.240425 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" containerName="neutron-db-sync" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.241860 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.257080 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.332871 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-564986d4bf-zhh4c"] Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.335251 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.338056 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.338326 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.338552 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-rdwdw" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.356780 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-564986d4bf-zhh4c"] Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.371998 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372068 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372258 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.372291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.473483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.473582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474065 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-httpd-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474179 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474223 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6chw4\" (UniqueName: \"kubernetes.io/projected/3ce663fe-b525-44ea-a6d4-33f6b2366f46-kube-api-access-6chw4\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474291 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474309 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474361 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-combined-ca-bundle\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474467 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474504 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.474843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.475215 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.475387 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.499926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"dnsmasq-dns-db9bd6759-f9ndc\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.565909 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576162 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6chw4\" (UniqueName: \"kubernetes.io/projected/3ce663fe-b525-44ea-a6d4-33f6b2366f46-kube-api-access-6chw4\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-combined-ca-bundle\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.576375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-httpd-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.582559 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-httpd-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.586549 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-combined-ca-bundle\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.596666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3ce663fe-b525-44ea-a6d4-33f6b2366f46-config\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.607071 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6chw4\" (UniqueName: \"kubernetes.io/projected/3ce663fe-b525-44ea-a6d4-33f6b2366f46-kube-api-access-6chw4\") pod \"neutron-564986d4bf-zhh4c\" (UID: \"3ce663fe-b525-44ea-a6d4-33f6b2366f46\") " pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:27 crc kubenswrapper[4907]: I0313 15:37:27.661529 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:28 crc kubenswrapper[4907]: I0313 15:37:28.146108 4907 scope.go:117] "RemoveContainer" containerID="bddac8c66ea0090f3f484cd37e970a8a28f550e7e44321363049a8343bf92762" Mar 13 15:37:28 crc kubenswrapper[4907]: I0313 15:37:28.153076 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:37:28 crc kubenswrapper[4907]: I0313 15:37:28.314125 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-564986d4bf-zhh4c"] Mar 13 15:37:28 crc kubenswrapper[4907]: W0313 15:37:28.320760 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ce663fe_b525_44ea_a6d4_33f6b2366f46.slice/crio-c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a WatchSource:0}: Error finding container c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a: Status 404 returned error can't find the container with id c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.026986 4907 generic.go:334] "Generic (PLEG): container finished" podID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerID="f003308b3b3b337edbb1d567791fd2916dac4c39389f2b18c7276a22c4799e5d" exitCode=0 Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.027210 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerDied","Data":"f003308b3b3b337edbb1d567791fd2916dac4c39389f2b18c7276a22c4799e5d"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.027385 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerStarted","Data":"89449f9bc1a34a373899c9c8f376372efe1d7f3851750d5a63cfd6c8a1ef4335"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030032 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564986d4bf-zhh4c" event={"ID":"3ce663fe-b525-44ea-a6d4-33f6b2366f46","Type":"ContainerStarted","Data":"4852d549b802ea4097d262c3246a3ef7bebacfab6f360b646e235ad408bb2f24"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030075 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564986d4bf-zhh4c" event={"ID":"3ce663fe-b525-44ea-a6d4-33f6b2366f46","Type":"ContainerStarted","Data":"6dc6b0fe7218c94a5f2886858f2d8f14a2cb91737d9ced5752f7e83123baa22d"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030085 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-564986d4bf-zhh4c" event={"ID":"3ce663fe-b525-44ea-a6d4-33f6b2366f46","Type":"ContainerStarted","Data":"c0afd6d8d5bedb0a51cafa4e726c26baac20fd293ee0e8cf9f73f3fe64e24c0a"} Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.030566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:29 crc kubenswrapper[4907]: I0313 15:37:29.093739 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-564986d4bf-zhh4c" podStartSLOduration=2.093715497 podStartE2EDuration="2.093715497s" podCreationTimestamp="2026-03-13 15:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:29.073162135 +0000 UTC m=+5547.972949824" watchObservedRunningTime="2026-03-13 15:37:29.093715497 +0000 UTC m=+5547.993503186" Mar 13 15:37:30 crc kubenswrapper[4907]: I0313 15:37:30.039319 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerStarted","Data":"181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba"} Mar 13 15:37:30 crc kubenswrapper[4907]: I0313 15:37:30.059768 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" podStartSLOduration=3.05975089 podStartE2EDuration="3.05975089s" podCreationTimestamp="2026-03-13 15:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:37:30.054390504 +0000 UTC m=+5548.954178203" watchObservedRunningTime="2026-03-13 15:37:30.05975089 +0000 UTC m=+5548.959538579" Mar 13 15:37:31 crc kubenswrapper[4907]: I0313 15:37:31.048192 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:37 crc kubenswrapper[4907]: I0313 15:37:37.567462 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:37:37 crc kubenswrapper[4907]: I0313 15:37:37.621817 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:37:37 crc kubenswrapper[4907]: I0313 15:37:37.622072 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" containerID="cri-o://c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" gracePeriod=10 Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.097122 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103817 4907 generic.go:334] "Generic (PLEG): container finished" podID="96134178-5784-41c4-bfb2-da280292c4ee" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" exitCode=0 Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerDied","Data":"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2"} Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" event={"ID":"96134178-5784-41c4-bfb2-da280292c4ee","Type":"ContainerDied","Data":"feefca8ed61869667fc95d91e0bb79d91fef75b784e7cbb8c5d167534ad176c0"} Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.103941 4907 scope.go:117] "RemoveContainer" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.104023 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd6bb59b7-mfqq8" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.144098 4907 scope.go:117] "RemoveContainer" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.163344 4907 scope.go:117] "RemoveContainer" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" Mar 13 15:37:38 crc kubenswrapper[4907]: E0313 15:37:38.163721 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2\": container with ID starting with c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2 not found: ID does not exist" containerID="c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.163756 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2"} err="failed to get container status \"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2\": rpc error: code = NotFound desc = could not find container \"c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2\": container with ID starting with c953880ea09d4c9e0288bb5fb9aab81c29c152c5094c1b1a2b88b549e9c408e2 not found: ID does not exist" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.163779 4907 scope.go:117] "RemoveContainer" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" Mar 13 15:37:38 crc kubenswrapper[4907]: E0313 15:37:38.164077 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c\": container with ID starting with 344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c not found: ID does not exist" containerID="344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.164102 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c"} err="failed to get container status \"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c\": rpc error: code = NotFound desc = could not find container \"344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c\": container with ID starting with 344113e606bf545fcc172c4d2e36f56d58e31f6999f405a725dfb2c2de7e6e6c not found: ID does not exist" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.167954 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168156 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.168200 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") pod \"96134178-5784-41c4-bfb2-da280292c4ee\" (UID: \"96134178-5784-41c4-bfb2-da280292c4ee\") " Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.175035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr" (OuterVolumeSpecName: "kube-api-access-6vmfr") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "kube-api-access-6vmfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.214138 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config" (OuterVolumeSpecName: "config") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.219087 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.219215 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.227057 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "96134178-5784-41c4-bfb2-da280292c4ee" (UID: "96134178-5784-41c4-bfb2-da280292c4ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270852 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270943 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vmfr\" (UniqueName: \"kubernetes.io/projected/96134178-5784-41c4-bfb2-da280292c4ee-kube-api-access-6vmfr\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270962 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270978 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.270989 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/96134178-5784-41c4-bfb2-da280292c4ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.450834 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:37:38 crc kubenswrapper[4907]: I0313 15:37:38.459484 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bd6bb59b7-mfqq8"] Mar 13 15:37:39 crc kubenswrapper[4907]: I0313 15:37:39.793666 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96134178-5784-41c4-bfb2-da280292c4ee" path="/var/lib/kubelet/pods/96134178-5784-41c4-bfb2-da280292c4ee/volumes" Mar 13 15:37:48 crc kubenswrapper[4907]: I0313 15:37:48.041737 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:37:48 crc kubenswrapper[4907]: I0313 15:37:48.042196 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.537137 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:37:53 crc kubenswrapper[4907]: E0313 15:37:53.538445 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.538470 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" Mar 13 15:37:53 crc kubenswrapper[4907]: E0313 15:37:53.538515 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="init" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.538530 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="init" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.538801 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="96134178-5784-41c4-bfb2-da280292c4ee" containerName="dnsmasq-dns" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.545116 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.548853 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.670957 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.671086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.671160 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772041 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772119 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772560 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.772714 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.797162 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"certified-operators-5k9nb\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:53 crc kubenswrapper[4907]: I0313 15:37:53.882203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:37:54 crc kubenswrapper[4907]: I0313 15:37:54.407392 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.249661 4907 generic.go:334] "Generic (PLEG): container finished" podID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" exitCode=0 Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.249762 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada"} Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.249991 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerStarted","Data":"6bf1cbbbca2dc292f4b7fb3a637b22871e833710be9542b6cacb01ec4e9d5066"} Mar 13 15:37:55 crc kubenswrapper[4907]: I0313 15:37:55.251406 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:37:56 crc kubenswrapper[4907]: I0313 15:37:56.258672 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerStarted","Data":"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c"} Mar 13 15:37:57 crc kubenswrapper[4907]: I0313 15:37:57.269982 4907 generic.go:334] "Generic (PLEG): container finished" podID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" exitCode=0 Mar 13 15:37:57 crc kubenswrapper[4907]: I0313 15:37:57.270040 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c"} Mar 13 15:37:57 crc kubenswrapper[4907]: I0313 15:37:57.672838 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-564986d4bf-zhh4c" Mar 13 15:37:58 crc kubenswrapper[4907]: I0313 15:37:58.282849 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerStarted","Data":"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a"} Mar 13 15:37:58 crc kubenswrapper[4907]: I0313 15:37:58.306030 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5k9nb" podStartSLOduration=2.492291323 podStartE2EDuration="5.306007688s" podCreationTimestamp="2026-03-13 15:37:53 +0000 UTC" firstStartedPulling="2026-03-13 15:37:55.251197866 +0000 UTC m=+5574.150985555" lastFinishedPulling="2026-03-13 15:37:58.064914241 +0000 UTC m=+5576.964701920" observedRunningTime="2026-03-13 15:37:58.296632803 +0000 UTC m=+5577.196420502" watchObservedRunningTime="2026-03-13 15:37:58.306007688 +0000 UTC m=+5577.205795377" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.139636 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.141069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.144468 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.144474 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.144667 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.148876 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.282443 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"auto-csr-approver-29556938-ptksb\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.384708 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"auto-csr-approver-29556938-ptksb\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.419948 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"auto-csr-approver-29556938-ptksb\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.460823 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:00 crc kubenswrapper[4907]: I0313 15:38:00.905840 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:38:01 crc kubenswrapper[4907]: I0313 15:38:01.305441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerStarted","Data":"e969e95aa6638b032cb7230dee7feb590118185e18e1ed4bf4c49a061143dfc1"} Mar 13 15:38:02 crc kubenswrapper[4907]: I0313 15:38:02.316852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerStarted","Data":"19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402"} Mar 13 15:38:02 crc kubenswrapper[4907]: I0313 15:38:02.337489 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556938-ptksb" podStartSLOduration=1.290600144 podStartE2EDuration="2.337460525s" podCreationTimestamp="2026-03-13 15:38:00 +0000 UTC" firstStartedPulling="2026-03-13 15:38:00.903940539 +0000 UTC m=+5579.803728218" lastFinishedPulling="2026-03-13 15:38:01.95080089 +0000 UTC m=+5580.850588599" observedRunningTime="2026-03-13 15:38:02.330208966 +0000 UTC m=+5581.229996655" watchObservedRunningTime="2026-03-13 15:38:02.337460525 +0000 UTC m=+5581.237248214" Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.325350 4907 generic.go:334] "Generic (PLEG): container finished" podID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerID="19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402" exitCode=0 Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.325577 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerDied","Data":"19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402"} Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.883369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.883424 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:03 crc kubenswrapper[4907]: I0313 15:38:03.944291 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.392276 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.448346 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.662387 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.771967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") pod \"6a5048c0-3add-450a-94c0-c068ceca72a4\" (UID: \"6a5048c0-3add-450a-94c0-c068ceca72a4\") " Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.788875 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7" (OuterVolumeSpecName: "kube-api-access-97bz7") pod "6a5048c0-3add-450a-94c0-c068ceca72a4" (UID: "6a5048c0-3add-450a-94c0-c068ceca72a4"). InnerVolumeSpecName "kube-api-access-97bz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.838680 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:38:04 crc kubenswrapper[4907]: E0313 15:38:04.839142 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerName="oc" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.839166 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerName="oc" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.839407 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" containerName="oc" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.840095 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.863919 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.875039 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97bz7\" (UniqueName: \"kubernetes.io/projected/6a5048c0-3add-450a-94c0-c068ceca72a4-kube-api-access-97bz7\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.909390 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.916221 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556932-fxhnm"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.939299 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.940647 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.942726 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.947416 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.976843 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:04 crc kubenswrapper[4907]: I0313 15:38:04.976953 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078041 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078264 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078390 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.078429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.079095 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.107406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"glance-db-create-kxfg2\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.158982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.179822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.179928 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.180841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.198116 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"glance-fd12-account-create-update-kxbcs\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.263599 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.342668 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556938-ptksb" event={"ID":"6a5048c0-3add-450a-94c0-c068ceca72a4","Type":"ContainerDied","Data":"e969e95aa6638b032cb7230dee7feb590118185e18e1ed4bf4c49a061143dfc1"} Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.342717 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e969e95aa6638b032cb7230dee7feb590118185e18e1ed4bf4c49a061143dfc1" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.342691 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556938-ptksb" Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.606015 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.705808 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:38:05 crc kubenswrapper[4907]: W0313 15:38:05.712328 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d6a6021_7cc4_42f9_848b_5bbfc3650ee7.slice/crio-83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495 WatchSource:0}: Error finding container 83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495: Status 404 returned error can't find the container with id 83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495 Mar 13 15:38:05 crc kubenswrapper[4907]: I0313 15:38:05.792669 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5efd747-3795-4922-9f81-3ef02dcb2f91" path="/var/lib/kubelet/pods/e5efd747-3795-4922-9f81-3ef02dcb2f91/volumes" Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.353040 4907 generic.go:334] "Generic (PLEG): container finished" podID="440d4795-da82-461c-b460-bdd5c271de4d" containerID="7414c52bdeec9a9b8ecea00c6ccd0ca4cd0ecf4d6c333567e9430b7214c0182e" exitCode=0 Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.353193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kxfg2" event={"ID":"440d4795-da82-461c-b460-bdd5c271de4d","Type":"ContainerDied","Data":"7414c52bdeec9a9b8ecea00c6ccd0ca4cd0ecf4d6c333567e9430b7214c0182e"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.353514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kxfg2" event={"ID":"440d4795-da82-461c-b460-bdd5c271de4d","Type":"ContainerStarted","Data":"7a689c1f01fe0ef145be6f24241adc9dc1ca896caf778fafb7281e1050d46874"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356091 4907 generic.go:334] "Generic (PLEG): container finished" podID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerID="38713faad1d72bf574ce7d6ae2bd740312997c935666c1fb81c567bd339dc0d0" exitCode=0 Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356158 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fd12-account-create-update-kxbcs" event={"ID":"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7","Type":"ContainerDied","Data":"38713faad1d72bf574ce7d6ae2bd740312997c935666c1fb81c567bd339dc0d0"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356252 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fd12-account-create-update-kxbcs" event={"ID":"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7","Type":"ContainerStarted","Data":"83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495"} Mar 13 15:38:06 crc kubenswrapper[4907]: I0313 15:38:06.356463 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5k9nb" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" containerID="cri-o://1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" gracePeriod=2 Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.312982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367649 4907 generic.go:334] "Generic (PLEG): container finished" podID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" exitCode=0 Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367695 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a"} Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367739 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k9nb" event={"ID":"46d919ab-084b-46cf-a55a-80060d0f89e5","Type":"ContainerDied","Data":"6bf1cbbbca2dc292f4b7fb3a637b22871e833710be9542b6cacb01ec4e9d5066"} Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k9nb" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.367755 4907 scope.go:117] "RemoveContainer" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.389499 4907 scope.go:117] "RemoveContainer" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.408513 4907 scope.go:117] "RemoveContainer" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.421322 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") pod \"46d919ab-084b-46cf-a55a-80060d0f89e5\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.421438 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") pod \"46d919ab-084b-46cf-a55a-80060d0f89e5\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.421549 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") pod \"46d919ab-084b-46cf-a55a-80060d0f89e5\" (UID: \"46d919ab-084b-46cf-a55a-80060d0f89e5\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.422696 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities" (OuterVolumeSpecName: "utilities") pod "46d919ab-084b-46cf-a55a-80060d0f89e5" (UID: "46d919ab-084b-46cf-a55a-80060d0f89e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.428755 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc" (OuterVolumeSpecName: "kube-api-access-pm6kc") pod "46d919ab-084b-46cf-a55a-80060d0f89e5" (UID: "46d919ab-084b-46cf-a55a-80060d0f89e5"). InnerVolumeSpecName "kube-api-access-pm6kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.490861 4907 scope.go:117] "RemoveContainer" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" Mar 13 15:38:07 crc kubenswrapper[4907]: E0313 15:38:07.492978 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a\": container with ID starting with 1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a not found: ID does not exist" containerID="1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493057 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a"} err="failed to get container status \"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a\": rpc error: code = NotFound desc = could not find container \"1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a\": container with ID starting with 1c7e727a5c90422b9d55f69cb636beca42802c6044e4986c0e34da16172b738a not found: ID does not exist" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493090 4907 scope.go:117] "RemoveContainer" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" Mar 13 15:38:07 crc kubenswrapper[4907]: E0313 15:38:07.493548 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c\": container with ID starting with b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c not found: ID does not exist" containerID="b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493578 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c"} err="failed to get container status \"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c\": rpc error: code = NotFound desc = could not find container \"b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c\": container with ID starting with b930f886e923a7aeedd0f47b3c8af96617296760b49fcc8a4cd563a5557bb45c not found: ID does not exist" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493597 4907 scope.go:117] "RemoveContainer" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" Mar 13 15:38:07 crc kubenswrapper[4907]: E0313 15:38:07.493920 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada\": container with ID starting with 07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada not found: ID does not exist" containerID="07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.493939 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada"} err="failed to get container status \"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada\": rpc error: code = NotFound desc = could not find container \"07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada\": container with ID starting with 07e6a739eb2cd151b3d0c8f4d016361ec920ef5cee0ff3c0cf3353c888330ada not found: ID does not exist" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.497831 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46d919ab-084b-46cf-a55a-80060d0f89e5" (UID: "46d919ab-084b-46cf-a55a-80060d0f89e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.524661 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm6kc\" (UniqueName: \"kubernetes.io/projected/46d919ab-084b-46cf-a55a-80060d0f89e5-kube-api-access-pm6kc\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.524705 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.524718 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d919ab-084b-46cf-a55a-80060d0f89e5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.765923 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.781264 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.810832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.810901 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5k9nb"] Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.829847 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") pod \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.830678 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" (UID: "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.933127 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") pod \"440d4795-da82-461c-b460-bdd5c271de4d\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.933261 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") pod \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\" (UID: \"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.933367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") pod \"440d4795-da82-461c-b460-bdd5c271de4d\" (UID: \"440d4795-da82-461c-b460-bdd5c271de4d\") " Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.934034 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.934034 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "440d4795-da82-461c-b460-bdd5c271de4d" (UID: "440d4795-da82-461c-b460-bdd5c271de4d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.938976 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2" (OuterVolumeSpecName: "kube-api-access-tzph2") pod "440d4795-da82-461c-b460-bdd5c271de4d" (UID: "440d4795-da82-461c-b460-bdd5c271de4d"). InnerVolumeSpecName "kube-api-access-tzph2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:07 crc kubenswrapper[4907]: I0313 15:38:07.940016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh" (OuterVolumeSpecName: "kube-api-access-99dgh") pod "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" (UID: "0d6a6021-7cc4-42f9-848b-5bbfc3650ee7"). InnerVolumeSpecName "kube-api-access-99dgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.034516 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzph2\" (UniqueName: \"kubernetes.io/projected/440d4795-da82-461c-b460-bdd5c271de4d-kube-api-access-tzph2\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.034555 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99dgh\" (UniqueName: \"kubernetes.io/projected/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7-kube-api-access-99dgh\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.034565 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/440d4795-da82-461c-b460-bdd5c271de4d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.379695 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kxfg2" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.379731 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kxfg2" event={"ID":"440d4795-da82-461c-b460-bdd5c271de4d","Type":"ContainerDied","Data":"7a689c1f01fe0ef145be6f24241adc9dc1ca896caf778fafb7281e1050d46874"} Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.379783 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a689c1f01fe0ef145be6f24241adc9dc1ca896caf778fafb7281e1050d46874" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.383066 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-fd12-account-create-update-kxbcs" event={"ID":"0d6a6021-7cc4-42f9-848b-5bbfc3650ee7","Type":"ContainerDied","Data":"83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495"} Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.383124 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83481b88571f9dd6758c4725b16318418738a3b1dc14c5db9de531f76d808495" Mar 13 15:38:08 crc kubenswrapper[4907]: I0313 15:38:08.383079 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-fd12-account-create-update-kxbcs" Mar 13 15:38:09 crc kubenswrapper[4907]: I0313 15:38:09.796586 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" path="/var/lib/kubelet/pods/46d919ab-084b-46cf-a55a-80060d0f89e5/volumes" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.062604 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.062969 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-utilities" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.062984 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-utilities" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.062998 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="440d4795-da82-461c-b460-bdd5c271de4d" containerName="mariadb-database-create" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063008 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="440d4795-da82-461c-b460-bdd5c271de4d" containerName="mariadb-database-create" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.063023 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerName="mariadb-account-create-update" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063032 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerName="mariadb-account-create-update" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.063049 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-content" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063056 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="extract-content" Mar 13 15:38:10 crc kubenswrapper[4907]: E0313 15:38:10.063067 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063074 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063232 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" containerName="mariadb-account-create-update" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063245 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="440d4795-da82-461c-b460-bdd5c271de4d" containerName="mariadb-database-create" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063252 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d919ab-084b-46cf-a55a-80060d0f89e5" containerName="registry-server" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.063793 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.065648 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5c589" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.065726 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069637 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069867 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.069947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.087525 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.171631 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.171720 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.171801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.172243 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.176680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.177123 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.178287 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.191036 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"glance-db-sync-tlb57\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:10 crc kubenswrapper[4907]: I0313 15:38:10.386579 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:11 crc kubenswrapper[4907]: I0313 15:38:11.017481 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:38:11 crc kubenswrapper[4907]: I0313 15:38:11.431840 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerStarted","Data":"3232e9eb498f3e8fa1c6c1dce83de18cc94e2bb946c98b463971e7acd7900377"} Mar 13 15:38:12 crc kubenswrapper[4907]: I0313 15:38:12.439953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerStarted","Data":"75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498"} Mar 13 15:38:12 crc kubenswrapper[4907]: I0313 15:38:12.461989 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tlb57" podStartSLOduration=2.461972963 podStartE2EDuration="2.461972963s" podCreationTimestamp="2026-03-13 15:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:12.453130921 +0000 UTC m=+5591.352918610" watchObservedRunningTime="2026-03-13 15:38:12.461972963 +0000 UTC m=+5591.361760642" Mar 13 15:38:15 crc kubenswrapper[4907]: I0313 15:38:15.474788 4907 generic.go:334] "Generic (PLEG): container finished" podID="13c6b679-96d2-407c-8b94-cd908d61945b" containerID="75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498" exitCode=0 Mar 13 15:38:15 crc kubenswrapper[4907]: I0313 15:38:15.474899 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerDied","Data":"75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498"} Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.853325 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890124 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890225 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890297 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.890362 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") pod \"13c6b679-96d2-407c-8b94-cd908d61945b\" (UID: \"13c6b679-96d2-407c-8b94-cd908d61945b\") " Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.899222 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2" (OuterVolumeSpecName: "kube-api-access-9t6d2") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "kube-api-access-9t6d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.901638 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.927806 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.940973 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data" (OuterVolumeSpecName: "config-data") pod "13c6b679-96d2-407c-8b94-cd908d61945b" (UID: "13c6b679-96d2-407c-8b94-cd908d61945b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991583 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t6d2\" (UniqueName: \"kubernetes.io/projected/13c6b679-96d2-407c-8b94-cd908d61945b-kube-api-access-9t6d2\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991626 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991640 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:16 crc kubenswrapper[4907]: I0313 15:38:16.991652 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13c6b679-96d2-407c-8b94-cd908d61945b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.489354 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tlb57" event={"ID":"13c6b679-96d2-407c-8b94-cd908d61945b","Type":"ContainerDied","Data":"3232e9eb498f3e8fa1c6c1dce83de18cc94e2bb946c98b463971e7acd7900377"} Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.489399 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3232e9eb498f3e8fa1c6c1dce83de18cc94e2bb946c98b463971e7acd7900377" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.489433 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tlb57" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.753089 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:17 crc kubenswrapper[4907]: E0313 15:38:17.753465 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" containerName="glance-db-sync" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.753483 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" containerName="glance-db-sync" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.753636 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" containerName="glance-db-sync" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.756779 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761300 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761393 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5c589" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761473 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.761761 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.774754 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.803970 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804045 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804070 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804104 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804121 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.804168 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.883766 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.887458 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.905816 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.905919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.906801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.906875 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907138 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907177 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907380 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907497 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.907704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.914074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.914767 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.917773 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.924766 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.934701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.940669 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"glance-default-external-api-0\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.973200 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.978217 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.986904 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 15:38:17 crc kubenswrapper[4907]: I0313 15:38:17.988289 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009429 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009556 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009655 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.009724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.042531 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.042606 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.042660 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.043399 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.043470 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4" gracePeriod=600 Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.080861 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112037 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112121 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112205 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112285 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112377 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112400 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112424 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112459 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112488 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.112515 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.113173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.113191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.117630 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.117793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.131169 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"dnsmasq-dns-746ff5d579-xk972\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.203654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233623 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233811 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233868 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233900 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.233938 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.234467 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.234930 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.240392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.242983 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.243081 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.243104 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.255680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"glance-default-internal-api-0\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.309842 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.451605 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510240 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4" exitCode=0 Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4"} Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510372 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5"} Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.510393 4907 scope.go:117] "RemoveContainer" containerID="592bf9f175f440e943369aab2abbb13ad2c26412cb0e0738c4a33a156b8a5596" Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.511752 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerStarted","Data":"4fce01782f7e90bf4e850aaa9469f5e98dfe763d1a2f2b92bdd79b3911f5c894"} Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.751363 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.819997 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:18 crc kubenswrapper[4907]: I0313 15:38:18.975486 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.534008 4907 generic.go:334] "Generic (PLEG): container finished" podID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" exitCode=0 Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.534326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerDied","Data":"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97"} Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.534413 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerStarted","Data":"ab14195333d30dbd417719105a07ed83adbe0a377eeede3562757ec9ea3f4db9"} Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.540608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerStarted","Data":"d459823cf8e274f3ff30a8db80d9e6b1c4d9208bd24b1d7094389f4a3c7c74dd"} Mar 13 15:38:19 crc kubenswrapper[4907]: I0313 15:38:19.553949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerStarted","Data":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.565067 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerStarted","Data":"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.566448 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.567760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerStarted","Data":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.567793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerStarted","Data":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.570587 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerStarted","Data":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.570729 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" containerID="cri-o://43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" gracePeriod=30 Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.571095 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" containerID="cri-o://11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" gracePeriod=30 Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.594302 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-746ff5d579-xk972" podStartSLOduration=3.5942807549999998 podStartE2EDuration="3.594280755s" podCreationTimestamp="2026-03-13 15:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:20.584397295 +0000 UTC m=+5599.484184984" watchObservedRunningTime="2026-03-13 15:38:20.594280755 +0000 UTC m=+5599.494068454" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.609380 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.609361427 podStartE2EDuration="3.609361427s" podCreationTimestamp="2026-03-13 15:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:20.605995294 +0000 UTC m=+5599.505782983" watchObservedRunningTime="2026-03-13 15:38:20.609361427 +0000 UTC m=+5599.509149116" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.624990 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.6249688129999997 podStartE2EDuration="3.624968813s" podCreationTimestamp="2026-03-13 15:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:20.624347296 +0000 UTC m=+5599.524134985" watchObservedRunningTime="2026-03-13 15:38:20.624968813 +0000 UTC m=+5599.524756502" Mar 13 15:38:20 crc kubenswrapper[4907]: I0313 15:38:20.787330 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.152211 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285356 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285427 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285471 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285516 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285715 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285750 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.285801 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") pod \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\" (UID: \"d38a5143-6b9b-4f3e-a41a-63668c7516ee\") " Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286089 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs" (OuterVolumeSpecName: "logs") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286632 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.286662 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38a5143-6b9b-4f3e-a41a-63668c7516ee-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.291936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7" (OuterVolumeSpecName: "kube-api-access-whgs7") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "kube-api-access-whgs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.302042 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts" (OuterVolumeSpecName: "scripts") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.305149 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph" (OuterVolumeSpecName: "ceph") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.315575 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.331821 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data" (OuterVolumeSpecName: "config-data") pod "d38a5143-6b9b-4f3e-a41a-63668c7516ee" (UID: "d38a5143-6b9b-4f3e-a41a-63668c7516ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389025 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389067 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389078 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389100 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whgs7\" (UniqueName: \"kubernetes.io/projected/d38a5143-6b9b-4f3e-a41a-63668c7516ee-kube-api-access-whgs7\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.389119 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38a5143-6b9b-4f3e-a41a-63668c7516ee-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578249 4907 generic.go:334] "Generic (PLEG): container finished" podID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" exitCode=0 Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578290 4907 generic.go:334] "Generic (PLEG): container finished" podID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" exitCode=143 Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578301 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578396 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerDied","Data":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578428 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerDied","Data":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d38a5143-6b9b-4f3e-a41a-63668c7516ee","Type":"ContainerDied","Data":"4fce01782f7e90bf4e850aaa9469f5e98dfe763d1a2f2b92bdd79b3911f5c894"} Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.578474 4907 scope.go:117] "RemoveContainer" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.610778 4907 scope.go:117] "RemoveContainer" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.618007 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.625130 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.631718 4907 scope.go:117] "RemoveContainer" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.632509 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": container with ID starting with 11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3 not found: ID does not exist" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632550 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} err="failed to get container status \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": rpc error: code = NotFound desc = could not find container \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": container with ID starting with 11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632579 4907 scope.go:117] "RemoveContainer" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.632870 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": container with ID starting with 43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564 not found: ID does not exist" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632936 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} err="failed to get container status \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": rpc error: code = NotFound desc = could not find container \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": container with ID starting with 43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.632951 4907 scope.go:117] "RemoveContainer" containerID="11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.633377 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3"} err="failed to get container status \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": rpc error: code = NotFound desc = could not find container \"11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3\": container with ID starting with 11786c2adfad844744530cf8121b7e49795de27cf3bb1b2d63adb0c8946c6ed3 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.633397 4907 scope.go:117] "RemoveContainer" containerID="43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.633646 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564"} err="failed to get container status \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": rpc error: code = NotFound desc = could not find container \"43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564\": container with ID starting with 43c65f745fbc9ed6cd5669f25b8b192f01f2ec6518de198f3e6e582ee13ac564 not found: ID does not exist" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.640441 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.640785 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.640800 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" Mar 13 15:38:21 crc kubenswrapper[4907]: E0313 15:38:21.640819 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.640828 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.641016 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-log" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.641032 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" containerName="glance-httpd" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.641875 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.643420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.676856 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798513 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798581 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798621 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798645 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798710 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798742 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.798774 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.806760 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d38a5143-6b9b-4f3e-a41a-63668c7516ee" path="/var/lib/kubelet/pods/d38a5143-6b9b-4f3e-a41a-63668c7516ee/volumes" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900220 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900299 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900477 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900516 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900544 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.900848 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.902310 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.920180 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.921450 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.922180 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.923420 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.924733 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"glance-default-external-api-0\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " pod="openstack/glance-default-external-api-0" Mar 13 15:38:21 crc kubenswrapper[4907]: I0313 15:38:21.966245 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.445748 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.589282 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerStarted","Data":"e6d5e941f6575995d025c2bf565c8da48222d14bbb2fbb1710f7cd79e1394fae"} Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.591294 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" containerID="cri-o://ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" gracePeriod=30 Mar 13 15:38:22 crc kubenswrapper[4907]: I0313 15:38:22.591406 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" containerID="cri-o://052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" gracePeriod=30 Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.293349 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438611 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438674 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438755 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438867 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.438951 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.439022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") pod \"3431e634-aed2-44d7-a5a2-251956514c4b\" (UID: \"3431e634-aed2-44d7-a5a2-251956514c4b\") " Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.441248 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.457188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph" (OuterVolumeSpecName: "ceph") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.458156 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts" (OuterVolumeSpecName: "scripts") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.459075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6" (OuterVolumeSpecName: "kube-api-access-vkjx6") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "kube-api-access-vkjx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.461122 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs" (OuterVolumeSpecName: "logs") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.501164 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541064 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541294 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541389 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkjx6\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-kube-api-access-vkjx6\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541461 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541552 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3431e634-aed2-44d7-a5a2-251956514c4b-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.541621 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/3431e634-aed2-44d7-a5a2-251956514c4b-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.575046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data" (OuterVolumeSpecName: "config-data") pod "3431e634-aed2-44d7-a5a2-251956514c4b" (UID: "3431e634-aed2-44d7-a5a2-251956514c4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601240 4907 generic.go:334] "Generic (PLEG): container finished" podID="3431e634-aed2-44d7-a5a2-251956514c4b" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" exitCode=0 Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601272 4907 generic.go:334] "Generic (PLEG): container finished" podID="3431e634-aed2-44d7-a5a2-251956514c4b" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" exitCode=143 Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601309 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerDied","Data":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerDied","Data":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601345 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3431e634-aed2-44d7-a5a2-251956514c4b","Type":"ContainerDied","Data":"d459823cf8e274f3ff30a8db80d9e6b1c4d9208bd24b1d7094389f4a3c7c74dd"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601365 4907 scope.go:117] "RemoveContainer" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.601513 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.605377 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerStarted","Data":"56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85"} Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.637254 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.643177 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3431e634-aed2-44d7-a5a2-251956514c4b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.644185 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.654709 4907 scope.go:117] "RemoveContainer" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.658382 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.659137 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659160 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.659174 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659325 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-httpd" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.659342 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" containerName="glance-log" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.660368 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.666848 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.675844 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.705059 4907 scope.go:117] "RemoveContainer" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.706410 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": container with ID starting with 052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830 not found: ID does not exist" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706448 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} err="failed to get container status \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": rpc error: code = NotFound desc = could not find container \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": container with ID starting with 052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706473 4907 scope.go:117] "RemoveContainer" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: E0313 15:38:23.706866 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": container with ID starting with ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6 not found: ID does not exist" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706902 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} err="failed to get container status \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": rpc error: code = NotFound desc = could not find container \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": container with ID starting with ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.706917 4907 scope.go:117] "RemoveContainer" containerID="052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.710225 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830"} err="failed to get container status \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": rpc error: code = NotFound desc = could not find container \"052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830\": container with ID starting with 052c7c2589da49f678cd5ba63c3a814c27749ed0800ee5b95fc97c6b92f33830 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.710264 4907 scope.go:117] "RemoveContainer" containerID="ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.711172 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6"} err="failed to get container status \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": rpc error: code = NotFound desc = could not find container \"ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6\": container with ID starting with ffa291e0576f78b774635c00fdaf08366bb2adfaf8a71181fa700bc295bbbdb6 not found: ID does not exist" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.798206 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3431e634-aed2-44d7-a5a2-251956514c4b" path="/var/lib/kubelet/pods/3431e634-aed2-44d7-a5a2-251956514c4b/volumes" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846585 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846911 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846933 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.846973 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.847171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.847482 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.847552 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.948984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949046 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949185 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949222 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.949263 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.950007 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.950010 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.954421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.954633 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.955064 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.955648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.967761 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"glance-default-internal-api-0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:38:23 crc kubenswrapper[4907]: I0313 15:38:23.986045 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:24 crc kubenswrapper[4907]: I0313 15:38:24.511495 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:38:24 crc kubenswrapper[4907]: W0313 15:38:24.515920 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod096d3713_ea4f_43d3_bfb9_9170e8958ed0.slice/crio-13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a WatchSource:0}: Error finding container 13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a: Status 404 returned error can't find the container with id 13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a Mar 13 15:38:24 crc kubenswrapper[4907]: I0313 15:38:24.617394 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerStarted","Data":"d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03"} Mar 13 15:38:24 crc kubenswrapper[4907]: I0313 15:38:24.619746 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerStarted","Data":"13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a"} Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.628978 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerStarted","Data":"27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f"} Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.629509 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerStarted","Data":"02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8"} Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.653168 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.653147532 podStartE2EDuration="2.653147532s" podCreationTimestamp="2026-03-13 15:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:25.645347748 +0000 UTC m=+5604.545135457" watchObservedRunningTime="2026-03-13 15:38:25.653147532 +0000 UTC m=+5604.552935221" Mar 13 15:38:25 crc kubenswrapper[4907]: I0313 15:38:25.654029 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.654019426 podStartE2EDuration="4.654019426s" podCreationTimestamp="2026-03-13 15:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:24.648193744 +0000 UTC m=+5603.547981433" watchObservedRunningTime="2026-03-13 15:38:25.654019426 +0000 UTC m=+5604.553807125" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.205018 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.224962 4907 scope.go:117] "RemoveContainer" containerID="44715012ac71de6a61b777f99c8ba8523d3ece17a5f46795a3605b96d66aee6e" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.275829 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.276075 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" containerID="cri-o://181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba" gracePeriod=10 Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.656991 4907 generic.go:334] "Generic (PLEG): container finished" podID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerID="181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba" exitCode=0 Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.657085 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerDied","Data":"181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba"} Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.760139 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935514 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935575 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935600 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.935820 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") pod \"b43d065a-571d-416f-96d3-f1713f3d3bdb\" (UID: \"b43d065a-571d-416f-96d3-f1713f3d3bdb\") " Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.944280 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh" (OuterVolumeSpecName: "kube-api-access-99knh") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "kube-api-access-99knh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.979250 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.980739 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config" (OuterVolumeSpecName: "config") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.983668 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:28 crc kubenswrapper[4907]: I0313 15:38:28.989362 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b43d065a-571d-416f-96d3-f1713f3d3bdb" (UID: "b43d065a-571d-416f-96d3-f1713f3d3bdb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037471 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037504 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037514 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037525 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b43d065a-571d-416f-96d3-f1713f3d3bdb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.037536 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99knh\" (UniqueName: \"kubernetes.io/projected/b43d065a-571d-416f-96d3-f1713f3d3bdb-kube-api-access-99knh\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.669154 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" event={"ID":"b43d065a-571d-416f-96d3-f1713f3d3bdb","Type":"ContainerDied","Data":"89449f9bc1a34a373899c9c8f376372efe1d7f3851750d5a63cfd6c8a1ef4335"} Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.669223 4907 scope.go:117] "RemoveContainer" containerID="181ece190f085ef12291d2abeb4a964937fb06d189ff825afc0a6770f3e932ba" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.669232 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db9bd6759-f9ndc" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.718766 4907 scope.go:117] "RemoveContainer" containerID="f003308b3b3b337edbb1d567791fd2916dac4c39389f2b18c7276a22c4799e5d" Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.727585 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.753142 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db9bd6759-f9ndc"] Mar 13 15:38:29 crc kubenswrapper[4907]: I0313 15:38:29.799945 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" path="/var/lib/kubelet/pods/b43d065a-571d-416f-96d3-f1713f3d3bdb/volumes" Mar 13 15:38:31 crc kubenswrapper[4907]: I0313 15:38:31.967608 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:38:31 crc kubenswrapper[4907]: I0313 15:38:31.967935 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.007147 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.016582 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.707307 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:38:32 crc kubenswrapper[4907]: I0313 15:38:32.707357 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:38:33 crc kubenswrapper[4907]: I0313 15:38:33.987071 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:33 crc kubenswrapper[4907]: I0313 15:38:33.987398 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.015229 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.045976 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.739382 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.739737 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.818126 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.818243 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 15:38:34 crc kubenswrapper[4907]: I0313 15:38:34.856566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:38:36 crc kubenswrapper[4907]: I0313 15:38:36.725794 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:36 crc kubenswrapper[4907]: I0313 15:38:36.730137 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.373243 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:38:46 crc kubenswrapper[4907]: E0313 15:38:46.374193 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="init" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.374208 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="init" Mar 13 15:38:46 crc kubenswrapper[4907]: E0313 15:38:46.374229 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.374235 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.374420 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b43d065a-571d-416f-96d3-f1713f3d3bdb" containerName="dnsmasq-dns" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.375013 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.384185 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.434145 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.434204 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.477707 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.479209 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.481479 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.489774 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535760 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535807 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.535995 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.537033 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.554643 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"placement-db-create-5svds\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.643811 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.643899 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.645193 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.665534 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"placement-c036-account-create-update-26jl8\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.692214 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:46 crc kubenswrapper[4907]: I0313 15:38:46.807146 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.156172 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.260694 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:38:47 crc kubenswrapper[4907]: W0313 15:38:47.267319 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e98ef6f_8765_4e4d_ab91_7dcdcb3f788d.slice/crio-4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c WatchSource:0}: Error finding container 4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c: Status 404 returned error can't find the container with id 4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.880942 4907 generic.go:334] "Generic (PLEG): container finished" podID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerID="dcfa157fb4134a5c5ee1865cf6f3b6a035a94e4eeb4f7a317b410b36548c37aa" exitCode=0 Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.881056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c036-account-create-update-26jl8" event={"ID":"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d","Type":"ContainerDied","Data":"dcfa157fb4134a5c5ee1865cf6f3b6a035a94e4eeb4f7a317b410b36548c37aa"} Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.881102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c036-account-create-update-26jl8" event={"ID":"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d","Type":"ContainerStarted","Data":"4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c"} Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.882773 4907 generic.go:334] "Generic (PLEG): container finished" podID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerID="48b0748e550ebe6b33142e46ea370e9b43cfd8df0c8482529bda2ee099868df0" exitCode=0 Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.882804 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5svds" event={"ID":"978c19e3-b11d-44ef-86d5-6b0674e488d1","Type":"ContainerDied","Data":"48b0748e550ebe6b33142e46ea370e9b43cfd8df0c8482529bda2ee099868df0"} Mar 13 15:38:47 crc kubenswrapper[4907]: I0313 15:38:47.882818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5svds" event={"ID":"978c19e3-b11d-44ef-86d5-6b0674e488d1","Type":"ContainerStarted","Data":"1868172bc20ad95b4823c0ddc65847d8216cf5daabbe1eacc5e6555c36492028"} Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.308800 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.316011 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") pod \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393347 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") pod \"978c19e3-b11d-44ef-86d5-6b0674e488d1\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393378 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") pod \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\" (UID: \"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.393552 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") pod \"978c19e3-b11d-44ef-86d5-6b0674e488d1\" (UID: \"978c19e3-b11d-44ef-86d5-6b0674e488d1\") " Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.394107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" (UID: "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.394303 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "978c19e3-b11d-44ef-86d5-6b0674e488d1" (UID: "978c19e3-b11d-44ef-86d5-6b0674e488d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.401121 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8" (OuterVolumeSpecName: "kube-api-access-9x5m8") pod "978c19e3-b11d-44ef-86d5-6b0674e488d1" (UID: "978c19e3-b11d-44ef-86d5-6b0674e488d1"). InnerVolumeSpecName "kube-api-access-9x5m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.401178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc" (OuterVolumeSpecName: "kube-api-access-pq9rc") pod "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" (UID: "0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d"). InnerVolumeSpecName "kube-api-access-pq9rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495376 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/978c19e3-b11d-44ef-86d5-6b0674e488d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495667 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495681 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x5m8\" (UniqueName: \"kubernetes.io/projected/978c19e3-b11d-44ef-86d5-6b0674e488d1-kube-api-access-9x5m8\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.495693 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq9rc\" (UniqueName: \"kubernetes.io/projected/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d-kube-api-access-pq9rc\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.900717 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c036-account-create-update-26jl8" event={"ID":"0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d","Type":"ContainerDied","Data":"4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c"} Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.900755 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c1252f178308a87af8cf89fb1bf3b3a1cf8b33192c8980c6d5095afcec1f78c" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.900765 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c036-account-create-update-26jl8" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.902305 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5svds" event={"ID":"978c19e3-b11d-44ef-86d5-6b0674e488d1","Type":"ContainerDied","Data":"1868172bc20ad95b4823c0ddc65847d8216cf5daabbe1eacc5e6555c36492028"} Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.902325 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1868172bc20ad95b4823c0ddc65847d8216cf5daabbe1eacc5e6555c36492028" Mar 13 15:38:49 crc kubenswrapper[4907]: I0313 15:38:49.902336 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5svds" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.845809 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:38:51 crc kubenswrapper[4907]: E0313 15:38:51.846555 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerName="mariadb-account-create-update" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerName="mariadb-account-create-update" Mar 13 15:38:51 crc kubenswrapper[4907]: E0313 15:38:51.846609 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerName="mariadb-database-create" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846618 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerName="mariadb-database-create" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846828 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" containerName="mariadb-account-create-update" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.846848 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" containerName="mariadb-database-create" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.849933 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.877908 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.910967 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.912819 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.917210 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6c87x" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.917260 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.917416 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.937987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938042 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938081 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938131 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938187 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938225 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938250 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938272 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.938295 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:51 crc kubenswrapper[4907]: I0313 15:38:51.940536 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049827 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049964 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.049986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050020 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050034 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050073 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.050092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.051172 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.051600 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.051800 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.052355 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.052905 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.057577 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.058025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.068964 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.069545 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"dnsmasq-dns-5dd7c58f95-579qq\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.070318 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"placement-db-sync-z7wl4\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.182362 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.254406 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.709624 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:38:52 crc kubenswrapper[4907]: W0313 15:38:52.711289 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5153eb33_98f7_4452_b4f5_751a6d49c54b.slice/crio-1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b WatchSource:0}: Error finding container 1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b: Status 404 returned error can't find the container with id 1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.807105 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.944667 4907 generic.go:334] "Generic (PLEG): container finished" podID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" exitCode=0 Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.945001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerDied","Data":"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78"} Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.945061 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerStarted","Data":"1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b"} Mar 13 15:38:52 crc kubenswrapper[4907]: I0313 15:38:52.949431 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerStarted","Data":"2f496bc28dfefe3747a49ec89a56b7e3d709569e49ca85c48f559a57416fa409"} Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.960571 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerStarted","Data":"10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0"} Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.964156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerStarted","Data":"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59"} Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.964343 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.982362 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-z7wl4" podStartSLOduration=2.982342353 podStartE2EDuration="2.982342353s" podCreationTimestamp="2026-03-13 15:38:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:53.973134831 +0000 UTC m=+5632.872922620" watchObservedRunningTime="2026-03-13 15:38:53.982342353 +0000 UTC m=+5632.882130052" Mar 13 15:38:53 crc kubenswrapper[4907]: I0313 15:38:53.994907 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" podStartSLOduration=2.994873926 podStartE2EDuration="2.994873926s" podCreationTimestamp="2026-03-13 15:38:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:53.992604083 +0000 UTC m=+5632.892391802" watchObservedRunningTime="2026-03-13 15:38:53.994873926 +0000 UTC m=+5632.894661625" Mar 13 15:38:54 crc kubenswrapper[4907]: I0313 15:38:54.977530 4907 generic.go:334] "Generic (PLEG): container finished" podID="a43a8ffb-4f15-4524-b750-517442c1f561" containerID="10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0" exitCode=0 Mar 13 15:38:54 crc kubenswrapper[4907]: I0313 15:38:54.977614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerDied","Data":"10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0"} Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.325459 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440383 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440619 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440720 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440798 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") pod \"a43a8ffb-4f15-4524-b750-517442c1f561\" (UID: \"a43a8ffb-4f15-4524-b750-517442c1f561\") " Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.440919 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs" (OuterVolumeSpecName: "logs") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.441620 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43a8ffb-4f15-4524-b750-517442c1f561-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.460084 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf" (OuterVolumeSpecName: "kube-api-access-wh4kf") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "kube-api-access-wh4kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.460213 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts" (OuterVolumeSpecName: "scripts") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.465140 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.475213 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data" (OuterVolumeSpecName: "config-data") pod "a43a8ffb-4f15-4524-b750-517442c1f561" (UID: "a43a8ffb-4f15-4524-b750-517442c1f561"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543414 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543458 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543470 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a43a8ffb-4f15-4524-b750-517442c1f561-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.543481 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh4kf\" (UniqueName: \"kubernetes.io/projected/a43a8ffb-4f15-4524-b750-517442c1f561-kube-api-access-wh4kf\") on node \"crc\" DevicePath \"\"" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.994659 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-z7wl4" event={"ID":"a43a8ffb-4f15-4524-b750-517442c1f561","Type":"ContainerDied","Data":"2f496bc28dfefe3747a49ec89a56b7e3d709569e49ca85c48f559a57416fa409"} Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.994697 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f496bc28dfefe3747a49ec89a56b7e3d709569e49ca85c48f559a57416fa409" Mar 13 15:38:56 crc kubenswrapper[4907]: I0313 15:38:56.994753 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-z7wl4" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.065268 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6575d7f87d-t4csd"] Mar 13 15:38:57 crc kubenswrapper[4907]: E0313 15:38:57.065651 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" containerName="placement-db-sync" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.065669 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" containerName="placement-db-sync" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.065834 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" containerName="placement-db-sync" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.066754 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.070850 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.071074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.071949 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6c87x" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.087211 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6575d7f87d-t4csd"] Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153618 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2dbv\" (UniqueName: \"kubernetes.io/projected/12e05309-4563-4b78-a598-e4e8ced9cc8d-kube-api-access-t2dbv\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153756 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-scripts\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153814 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-combined-ca-bundle\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153849 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-config-data\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.153870 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e05309-4563-4b78-a598-e4e8ced9cc8d-logs\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-combined-ca-bundle\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-config-data\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e05309-4563-4b78-a598-e4e8ced9cc8d-logs\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255278 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2dbv\" (UniqueName: \"kubernetes.io/projected/12e05309-4563-4b78-a598-e4e8ced9cc8d-kube-api-access-t2dbv\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255326 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-scripts\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.255714 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12e05309-4563-4b78-a598-e4e8ced9cc8d-logs\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.259491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-combined-ca-bundle\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.260822 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-scripts\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.269955 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e05309-4563-4b78-a598-e4e8ced9cc8d-config-data\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.270683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2dbv\" (UniqueName: \"kubernetes.io/projected/12e05309-4563-4b78-a598-e4e8ced9cc8d-kube-api-access-t2dbv\") pod \"placement-6575d7f87d-t4csd\" (UID: \"12e05309-4563-4b78-a598-e4e8ced9cc8d\") " pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.391707 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:57 crc kubenswrapper[4907]: I0313 15:38:57.849214 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6575d7f87d-t4csd"] Mar 13 15:38:57 crc kubenswrapper[4907]: W0313 15:38:57.855238 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e05309_4563_4b78_a598_e4e8ced9cc8d.slice/crio-3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0 WatchSource:0}: Error finding container 3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0: Status 404 returned error can't find the container with id 3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0 Mar 13 15:38:58 crc kubenswrapper[4907]: I0313 15:38:58.007760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6575d7f87d-t4csd" event={"ID":"12e05309-4563-4b78-a598-e4e8ced9cc8d","Type":"ContainerStarted","Data":"3ccd101742905092f11d499e1eb4ca102c9c0edc448d3af36ebab529b04074f0"} Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.018050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6575d7f87d-t4csd" event={"ID":"12e05309-4563-4b78-a598-e4e8ced9cc8d","Type":"ContainerStarted","Data":"c86e60fc008ccf1955bb3b2212ab51aaed4913a4393b406fcbdfd6f106d7909f"} Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.019966 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6575d7f87d-t4csd" event={"ID":"12e05309-4563-4b78-a598-e4e8ced9cc8d","Type":"ContainerStarted","Data":"84c89fa87368ab0d7170f4175671afe4126cfa61e8fd73f4ce834716a16cef38"} Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.019996 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:38:59 crc kubenswrapper[4907]: I0313 15:38:59.037642 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6575d7f87d-t4csd" podStartSLOduration=2.037616873 podStartE2EDuration="2.037616873s" podCreationTimestamp="2026-03-13 15:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:38:59.033059197 +0000 UTC m=+5637.932846926" watchObservedRunningTime="2026-03-13 15:38:59.037616873 +0000 UTC m=+5637.937404592" Mar 13 15:39:00 crc kubenswrapper[4907]: I0313 15:39:00.027711 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.184095 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.243287 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.252590 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-746ff5d579-xk972" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" containerID="cri-o://be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" gracePeriod=10 Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.691630 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767547 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767699 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767896 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.767956 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") pod \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\" (UID: \"4cf4ac52-db94-4e0c-a990-c2c6c64f3167\") " Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.774316 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s" (OuterVolumeSpecName: "kube-api-access-4h66s") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "kube-api-access-4h66s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.811426 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config" (OuterVolumeSpecName: "config") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.812132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.815457 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.818491 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4cf4ac52-db94-4e0c-a990-c2c6c64f3167" (UID: "4cf4ac52-db94-4e0c-a990-c2c6c64f3167"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869301 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h66s\" (UniqueName: \"kubernetes.io/projected/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-kube-api-access-4h66s\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869335 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869348 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869357 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:02 crc kubenswrapper[4907]: I0313 15:39:02.869366 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf4ac52-db94-4e0c-a990-c2c6c64f3167-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066719 4907 generic.go:334] "Generic (PLEG): container finished" podID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" exitCode=0 Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066764 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-746ff5d579-xk972" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerDied","Data":"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e"} Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066817 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-746ff5d579-xk972" event={"ID":"4cf4ac52-db94-4e0c-a990-c2c6c64f3167","Type":"ContainerDied","Data":"ab14195333d30dbd417719105a07ed83adbe0a377eeede3562757ec9ea3f4db9"} Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.066846 4907 scope.go:117] "RemoveContainer" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.086196 4907 scope.go:117] "RemoveContainer" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.101047 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.107590 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-746ff5d579-xk972"] Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.111397 4907 scope.go:117] "RemoveContainer" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" Mar 13 15:39:03 crc kubenswrapper[4907]: E0313 15:39:03.111863 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e\": container with ID starting with be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e not found: ID does not exist" containerID="be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.111927 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e"} err="failed to get container status \"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e\": rpc error: code = NotFound desc = could not find container \"be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e\": container with ID starting with be908740b0d62c157a0b40660ddc6349bfde1dfa92e3ea3447a7f72e7e36459e not found: ID does not exist" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.111958 4907 scope.go:117] "RemoveContainer" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" Mar 13 15:39:03 crc kubenswrapper[4907]: E0313 15:39:03.112305 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97\": container with ID starting with 2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97 not found: ID does not exist" containerID="2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.112352 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97"} err="failed to get container status \"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97\": rpc error: code = NotFound desc = could not find container \"2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97\": container with ID starting with 2d77589298c0e5cd48847476ea24e4e7c3f37cb59a1ce9b8eb1bd699e22f7f97 not found: ID does not exist" Mar 13 15:39:03 crc kubenswrapper[4907]: I0313 15:39:03.796367 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" path="/var/lib/kubelet/pods/4cf4ac52-db94-4e0c-a990-c2c6c64f3167/volumes" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.387644 4907 scope.go:117] "RemoveContainer" containerID="3c35c6992ec9b87a4cfcf6f80130463db4ea23212e37b3ceef2bd0131f1b10b1" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.392247 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.405083 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6575d7f87d-t4csd" Mar 13 15:39:28 crc kubenswrapper[4907]: I0313 15:39:28.406056 4907 scope.go:117] "RemoveContainer" containerID="2c2f32fc7b86119b10010b256cedf2bb75451fdf3804d5c1b2c61e62707d4b83" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.727654 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:39:51 crc kubenswrapper[4907]: E0313 15:39:51.728570 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.728585 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" Mar 13 15:39:51 crc kubenswrapper[4907]: E0313 15:39:51.728612 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="init" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.728638 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="init" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.728836 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cf4ac52-db94-4e0c-a990-c2c6c64f3167" containerName="dnsmasq-dns" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.729543 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.738980 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.827434 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.828773 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.836960 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.855358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.855416 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.951654 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.953076 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956752 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956803 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956835 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.956935 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.957530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.959482 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.960541 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.962175 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.970505 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:39:51 crc kubenswrapper[4907]: I0313 15:39:51.983793 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.012514 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"nova-api-db-create-5r4tb\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.051240 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065142 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065260 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.065308 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.066343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.110801 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"nova-cell0-db-create-brbzj\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.144621 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.157743 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.159852 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.165204 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166614 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166720 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166745 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.166793 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.167481 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.168315 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.169684 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.192701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"nova-cell1-db-create-7djpm\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.193190 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"nova-api-62d0-account-create-update-vf8sr\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.268523 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.268583 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.271360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.284448 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.345469 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.348783 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.351010 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.354759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.370231 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.370489 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.371500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.386557 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"nova-cell0-356d-account-create-update-kb8lr\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.473378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.473491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.574321 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.575781 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.575972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.577260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.592897 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"nova-cell1-5a99-account-create-update-nkg4x\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.629768 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.685085 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:52 crc kubenswrapper[4907]: W0313 15:39:52.734997 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18fa9304_50a2_45cf_b983_9d459f0dcb5f.slice/crio-e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653 WatchSource:0}: Error finding container e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653: Status 404 returned error can't find the container with id e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653 Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.737407 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.847360 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:39:52 crc kubenswrapper[4907]: W0313 15:39:52.859755 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2c52fb5_c4ba_4c24_95af_ce75cd609126.slice/crio-0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90 WatchSource:0}: Error finding container 0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90: Status 404 returned error can't find the container with id 0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90 Mar 13 15:39:52 crc kubenswrapper[4907]: I0313 15:39:52.869758 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.089899 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:39:53 crc kubenswrapper[4907]: W0313 15:39:53.131412 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb66093c1_e494_4975_b8d1_217844060204.slice/crio-0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7 WatchSource:0}: Error finding container 0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7: Status 404 returned error can't find the container with id 0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.203682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:39:53 crc kubenswrapper[4907]: W0313 15:39:53.225181 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40763271_36d9_4bd2_8ae8_82140648fcf4.slice/crio-543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8 WatchSource:0}: Error finding container 543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8: Status 404 returned error can't find the container with id 543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.508675 4907 generic.go:334] "Generic (PLEG): container finished" podID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerID="f18dc1660972c456737f08b933ea252844a2bf45813b5a1dae54e428ebc166a5" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.508742 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-brbzj" event={"ID":"18fa9304-50a2-45cf-b983-9d459f0dcb5f","Type":"ContainerDied","Data":"f18dc1660972c456737f08b933ea252844a2bf45813b5a1dae54e428ebc166a5"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.508769 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-brbzj" event={"ID":"18fa9304-50a2-45cf-b983-9d459f0dcb5f","Type":"ContainerStarted","Data":"e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.510938 4907 generic.go:334] "Generic (PLEG): container finished" podID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerID="c18542c3b44a6f6eb28a26dfdcfcff6b4ceb92d4f5c1a2a581f9880ab136aa90" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.510988 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-62d0-account-create-update-vf8sr" event={"ID":"94ba887c-f88a-48e7-b4bc-2f171d21fa47","Type":"ContainerDied","Data":"c18542c3b44a6f6eb28a26dfdcfcff6b4ceb92d4f5c1a2a581f9880ab136aa90"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.511004 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-62d0-account-create-update-vf8sr" event={"ID":"94ba887c-f88a-48e7-b4bc-2f171d21fa47","Type":"ContainerStarted","Data":"4a40897775cd1b3c4b3543b2070f41a576ca4e865bdb987c81c613efd9443cc6"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.512773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" event={"ID":"b66093c1-e494-4975-b8d1-217844060204","Type":"ContainerStarted","Data":"0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.514550 4907 generic.go:334] "Generic (PLEG): container finished" podID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerID="d3e4c590d97414f576974cec65543e931ac60a3bcf12387dbe9df0a1a39e2c28" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.514602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5r4tb" event={"ID":"e5aab76f-4bcc-47be-989d-7d9d6ecf371b","Type":"ContainerDied","Data":"d3e4c590d97414f576974cec65543e931ac60a3bcf12387dbe9df0a1a39e2c28"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.514621 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5r4tb" event={"ID":"e5aab76f-4bcc-47be-989d-7d9d6ecf371b","Type":"ContainerStarted","Data":"dd3f82f5adfcd093fbfe4c95a182789207aea1df8bcb504fdcb2fe97dbfbdb96"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.517154 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" event={"ID":"40763271-36d9-4bd2-8ae8-82140648fcf4","Type":"ContainerStarted","Data":"543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.519103 4907 generic.go:334] "Generic (PLEG): container finished" podID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerID="9d15efc6444fb5cd991e1e807468f40b653afac37426c92c9f3e8218e37197cf" exitCode=0 Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.519149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7djpm" event={"ID":"c2c52fb5-c4ba-4c24-95af-ce75cd609126","Type":"ContainerDied","Data":"9d15efc6444fb5cd991e1e807468f40b653afac37426c92c9f3e8218e37197cf"} Mar 13 15:39:53 crc kubenswrapper[4907]: I0313 15:39:53.519170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7djpm" event={"ID":"c2c52fb5-c4ba-4c24-95af-ce75cd609126","Type":"ContainerStarted","Data":"0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90"} Mar 13 15:39:53 crc kubenswrapper[4907]: E0313 15:39:53.912491 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb66093c1_e494_4975_b8d1_217844060204.slice/crio-conmon-c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.531553 4907 generic.go:334] "Generic (PLEG): container finished" podID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerID="0948541af6cc99af21bd6ee88e49185bc604c42dbf31f8661316b429337a2c7f" exitCode=0 Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.531649 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" event={"ID":"40763271-36d9-4bd2-8ae8-82140648fcf4","Type":"ContainerDied","Data":"0948541af6cc99af21bd6ee88e49185bc604c42dbf31f8661316b429337a2c7f"} Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.535573 4907 generic.go:334] "Generic (PLEG): container finished" podID="b66093c1-e494-4975-b8d1-217844060204" containerID="c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96" exitCode=0 Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.535643 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" event={"ID":"b66093c1-e494-4975-b8d1-217844060204","Type":"ContainerDied","Data":"c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96"} Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.874190 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.947123 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") pod \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.947218 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") pod \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\" (UID: \"e5aab76f-4bcc-47be-989d-7d9d6ecf371b\") " Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.947987 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5aab76f-4bcc-47be-989d-7d9d6ecf371b" (UID: "e5aab76f-4bcc-47be-989d-7d9d6ecf371b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.949488 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:54 crc kubenswrapper[4907]: I0313 15:39:54.953155 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc" (OuterVolumeSpecName: "kube-api-access-b4jdc") pod "e5aab76f-4bcc-47be-989d-7d9d6ecf371b" (UID: "e5aab76f-4bcc-47be-989d-7d9d6ecf371b"). InnerVolumeSpecName "kube-api-access-b4jdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.041663 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.046438 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.051275 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.051802 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4jdc\" (UniqueName: \"kubernetes.io/projected/e5aab76f-4bcc-47be-989d-7d9d6ecf371b-kube-api-access-b4jdc\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.152778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") pod \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.152928 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") pod \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.152955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") pod \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\" (UID: \"c2c52fb5-c4ba-4c24-95af-ce75cd609126\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153002 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") pod \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153063 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") pod \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\" (UID: \"94ba887c-f88a-48e7-b4bc-2f171d21fa47\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153111 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") pod \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\" (UID: \"18fa9304-50a2-45cf-b983-9d459f0dcb5f\") " Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153385 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2c52fb5-c4ba-4c24-95af-ce75cd609126" (UID: "c2c52fb5-c4ba-4c24-95af-ce75cd609126"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153455 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18fa9304-50a2-45cf-b983-9d459f0dcb5f" (UID: "18fa9304-50a2-45cf-b983-9d459f0dcb5f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153478 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "94ba887c-f88a-48e7-b4bc-2f171d21fa47" (UID: "94ba887c-f88a-48e7-b4bc-2f171d21fa47"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153895 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2c52fb5-c4ba-4c24-95af-ce75cd609126-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153914 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18fa9304-50a2-45cf-b983-9d459f0dcb5f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.153923 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94ba887c-f88a-48e7-b4bc-2f171d21fa47-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.156066 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp" (OuterVolumeSpecName: "kube-api-access-vrtdp") pod "18fa9304-50a2-45cf-b983-9d459f0dcb5f" (UID: "18fa9304-50a2-45cf-b983-9d459f0dcb5f"). InnerVolumeSpecName "kube-api-access-vrtdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.156114 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b" (OuterVolumeSpecName: "kube-api-access-64c2b") pod "94ba887c-f88a-48e7-b4bc-2f171d21fa47" (UID: "94ba887c-f88a-48e7-b4bc-2f171d21fa47"). InnerVolumeSpecName "kube-api-access-64c2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.157783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq" (OuterVolumeSpecName: "kube-api-access-lxptq") pod "c2c52fb5-c4ba-4c24-95af-ce75cd609126" (UID: "c2c52fb5-c4ba-4c24-95af-ce75cd609126"). InnerVolumeSpecName "kube-api-access-lxptq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.255844 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64c2b\" (UniqueName: \"kubernetes.io/projected/94ba887c-f88a-48e7-b4bc-2f171d21fa47-kube-api-access-64c2b\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.255942 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxptq\" (UniqueName: \"kubernetes.io/projected/c2c52fb5-c4ba-4c24-95af-ce75cd609126-kube-api-access-lxptq\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.255953 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrtdp\" (UniqueName: \"kubernetes.io/projected/18fa9304-50a2-45cf-b983-9d459f0dcb5f-kube-api-access-vrtdp\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.549233 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5r4tb" event={"ID":"e5aab76f-4bcc-47be-989d-7d9d6ecf371b","Type":"ContainerDied","Data":"dd3f82f5adfcd093fbfe4c95a182789207aea1df8bcb504fdcb2fe97dbfbdb96"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.549287 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd3f82f5adfcd093fbfe4c95a182789207aea1df8bcb504fdcb2fe97dbfbdb96" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.549351 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5r4tb" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.555478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7djpm" event={"ID":"c2c52fb5-c4ba-4c24-95af-ce75cd609126","Type":"ContainerDied","Data":"0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.555540 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e23da64c584ff8030a07246c197bc4df16cf45b1b7b65a2a1e9a6373f844c90" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.555626 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7djpm" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.561122 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-brbzj" event={"ID":"18fa9304-50a2-45cf-b983-9d459f0dcb5f","Type":"ContainerDied","Data":"e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.561168 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2b380c63ddd685de31e5a747984baea0335628ccbc6b3f1af44debb13f13653" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.561224 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-brbzj" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.564310 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-62d0-account-create-update-vf8sr" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.564968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-62d0-account-create-update-vf8sr" event={"ID":"94ba887c-f88a-48e7-b4bc-2f171d21fa47","Type":"ContainerDied","Data":"4a40897775cd1b3c4b3543b2070f41a576ca4e865bdb987c81c613efd9443cc6"} Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.565024 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a40897775cd1b3c4b3543b2070f41a576ca4e865bdb987c81c613efd9443cc6" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.925042 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:55 crc kubenswrapper[4907]: I0313 15:39:55.930684 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.069865 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") pod \"40763271-36d9-4bd2-8ae8-82140648fcf4\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.070202 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") pod \"b66093c1-e494-4975-b8d1-217844060204\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.070252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") pod \"40763271-36d9-4bd2-8ae8-82140648fcf4\" (UID: \"40763271-36d9-4bd2-8ae8-82140648fcf4\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.070548 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") pod \"b66093c1-e494-4975-b8d1-217844060204\" (UID: \"b66093c1-e494-4975-b8d1-217844060204\") " Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.073222 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b66093c1-e494-4975-b8d1-217844060204" (UID: "b66093c1-e494-4975-b8d1-217844060204"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.073543 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40763271-36d9-4bd2-8ae8-82140648fcf4" (UID: "40763271-36d9-4bd2-8ae8-82140648fcf4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.081004 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl" (OuterVolumeSpecName: "kube-api-access-pkzvl") pod "40763271-36d9-4bd2-8ae8-82140648fcf4" (UID: "40763271-36d9-4bd2-8ae8-82140648fcf4"). InnerVolumeSpecName "kube-api-access-pkzvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.087300 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9" (OuterVolumeSpecName: "kube-api-access-wcjg9") pod "b66093c1-e494-4975-b8d1-217844060204" (UID: "b66093c1-e494-4975-b8d1-217844060204"). InnerVolumeSpecName "kube-api-access-wcjg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173590 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcjg9\" (UniqueName: \"kubernetes.io/projected/b66093c1-e494-4975-b8d1-217844060204-kube-api-access-wcjg9\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173640 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40763271-36d9-4bd2-8ae8-82140648fcf4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173653 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b66093c1-e494-4975-b8d1-217844060204-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.173665 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkzvl\" (UniqueName: \"kubernetes.io/projected/40763271-36d9-4bd2-8ae8-82140648fcf4-kube-api-access-pkzvl\") on node \"crc\" DevicePath \"\"" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.575802 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.575825 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-356d-account-create-update-kb8lr" event={"ID":"b66093c1-e494-4975-b8d1-217844060204","Type":"ContainerDied","Data":"0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7"} Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.575937 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bf11b5c16f026ec4133b29e8d2e6b391b0da007fcd4acde47500c6e6fdc42e7" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.577682 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" event={"ID":"40763271-36d9-4bd2-8ae8-82140648fcf4","Type":"ContainerDied","Data":"543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8"} Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.577731 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="543db5418a330957dba2de6b68ab5a198dc372131f880d599abcd7de199ca1e8" Mar 13 15:39:56 crc kubenswrapper[4907]: I0313 15:39:56.577790 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5a99-account-create-update-nkg4x" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330285 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330637 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330650 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330675 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b66093c1-e494-4975-b8d1-217844060204" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330681 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b66093c1-e494-4975-b8d1-217844060204" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330689 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330703 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330709 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330721 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330727 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: E0313 15:39:57.330746 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330752 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330910 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330923 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330935 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330949 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b66093c1-e494-4975-b8d1-217844060204" containerName="mariadb-account-create-update" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330960 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.330970 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" containerName="mariadb-database-create" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.331826 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.334751 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.334800 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fmnpb" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.335023 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.343414 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394143 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394319 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.394340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495292 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495400 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.495434 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.499194 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.499846 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.501031 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.517152 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"nova-cell0-conductor-db-sync-mwt47\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:57 crc kubenswrapper[4907]: I0313 15:39:57.648950 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.086611 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:39:58 crc kubenswrapper[4907]: W0313 15:39:58.092382 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0336c3c0_cc18_4bb3_afe3_7e2a0232c889.slice/crio-fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92 WatchSource:0}: Error finding container fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92: Status 404 returned error can't find the container with id fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92 Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.596045 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerStarted","Data":"e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186"} Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.596343 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerStarted","Data":"fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92"} Mar 13 15:39:58 crc kubenswrapper[4907]: I0313 15:39:58.624031 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-mwt47" podStartSLOduration=1.6240110250000002 podStartE2EDuration="1.624011025s" podCreationTimestamp="2026-03-13 15:39:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:39:58.612334395 +0000 UTC m=+5697.512122084" watchObservedRunningTime="2026-03-13 15:39:58.624011025 +0000 UTC m=+5697.523798714" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.137469 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.139224 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.145583 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.145635 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.146041 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.147306 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.252269 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"auto-csr-approver-29556940-5rmt4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.354023 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"auto-csr-approver-29556940-5rmt4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.371090 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"auto-csr-approver-29556940-5rmt4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.465977 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:00 crc kubenswrapper[4907]: I0313 15:40:00.889875 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:40:00 crc kubenswrapper[4907]: W0313 15:40:00.895296 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72d20fdf_a093_459f_9328_0549ef7b48c4.slice/crio-d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4 WatchSource:0}: Error finding container d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4: Status 404 returned error can't find the container with id d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4 Mar 13 15:40:01 crc kubenswrapper[4907]: I0313 15:40:01.629184 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" event={"ID":"72d20fdf-a093-459f-9328-0549ef7b48c4","Type":"ContainerStarted","Data":"d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4"} Mar 13 15:40:02 crc kubenswrapper[4907]: I0313 15:40:02.640145 4907 generic.go:334] "Generic (PLEG): container finished" podID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerID="3f861dd2c3ff8fa5314ce9c35d45445320ad82d6381b4138837ad16c289319f5" exitCode=0 Mar 13 15:40:02 crc kubenswrapper[4907]: I0313 15:40:02.640205 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" event={"ID":"72d20fdf-a093-459f-9328-0549ef7b48c4","Type":"ContainerDied","Data":"3f861dd2c3ff8fa5314ce9c35d45445320ad82d6381b4138837ad16c289319f5"} Mar 13 15:40:03 crc kubenswrapper[4907]: I0313 15:40:03.651224 4907 generic.go:334] "Generic (PLEG): container finished" podID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerID="e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186" exitCode=0 Mar 13 15:40:03 crc kubenswrapper[4907]: I0313 15:40:03.651302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerDied","Data":"e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186"} Mar 13 15:40:03 crc kubenswrapper[4907]: I0313 15:40:03.949648 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.014991 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") pod \"72d20fdf-a093-459f-9328-0549ef7b48c4\" (UID: \"72d20fdf-a093-459f-9328-0549ef7b48c4\") " Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.020922 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj" (OuterVolumeSpecName: "kube-api-access-v7mcj") pod "72d20fdf-a093-459f-9328-0549ef7b48c4" (UID: "72d20fdf-a093-459f-9328-0549ef7b48c4"). InnerVolumeSpecName "kube-api-access-v7mcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.116565 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7mcj\" (UniqueName: \"kubernetes.io/projected/72d20fdf-a093-459f-9328-0549ef7b48c4-kube-api-access-v7mcj\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.660437 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" event={"ID":"72d20fdf-a093-459f-9328-0549ef7b48c4","Type":"ContainerDied","Data":"d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4"} Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.660480 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2776116aa3026d03d3e50e4d891cc0f1286bb990c15bcd8292e00f25ab105e4" Mar 13 15:40:04 crc kubenswrapper[4907]: I0313 15:40:04.660489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556940-5rmt4" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.016668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.020466 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.027713 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556934-wf7zr"] Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.136411 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.136724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.136830 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.137041 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") pod \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\" (UID: \"0336c3c0-cc18-4bb3-afe3-7e2a0232c889\") " Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.141736 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts" (OuterVolumeSpecName: "scripts") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.142039 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t" (OuterVolumeSpecName: "kube-api-access-rxd2t") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "kube-api-access-rxd2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.163387 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.165593 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data" (OuterVolumeSpecName: "config-data") pod "0336c3c0-cc18-4bb3-afe3-7e2a0232c889" (UID: "0336c3c0-cc18-4bb3-afe3-7e2a0232c889"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238921 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238968 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxd2t\" (UniqueName: \"kubernetes.io/projected/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-kube-api-access-rxd2t\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238982 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.238993 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0336c3c0-cc18-4bb3-afe3-7e2a0232c889-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.669775 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mwt47" event={"ID":"0336c3c0-cc18-4bb3-afe3-7e2a0232c889","Type":"ContainerDied","Data":"fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92"} Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.669808 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd3776051992ad0e194814f0af106e536e760a3d9887d7bff525ead01fdacb92" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.669896 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mwt47" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.741176 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:40:05 crc kubenswrapper[4907]: E0313 15:40:05.742173 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerName="nova-cell0-conductor-db-sync" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742206 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerName="nova-cell0-conductor-db-sync" Mar 13 15:40:05 crc kubenswrapper[4907]: E0313 15:40:05.742269 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerName="oc" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742281 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerName="oc" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742521 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" containerName="oc" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.742563 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" containerName="nova-cell0-conductor-db-sync" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.743481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.745613 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fmnpb" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.762115 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.765348 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.805568 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="041cab28-d0e1-4ec9-a0cb-92176be369aa" path="/var/lib/kubelet/pods/041cab28-d0e1-4ec9-a0cb-92176be369aa/volumes" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.847313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.847400 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.847717 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.949460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.949535 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.949564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.953373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.953962 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:05 crc kubenswrapper[4907]: I0313 15:40:05.967506 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"nova-cell0-conductor-0\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.069834 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.490604 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.678927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerStarted","Data":"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34"} Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.678977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerStarted","Data":"78d4b76c3a1fd63fabbe50a18b7db60eba6e1e2f95e60b44b880d052c99e1044"} Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.679073 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:06 crc kubenswrapper[4907]: I0313 15:40:06.703740 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.703722275 podStartE2EDuration="1.703722275s" podCreationTimestamp="2026-03-13 15:40:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:06.694908715 +0000 UTC m=+5705.594696404" watchObservedRunningTime="2026-03-13 15:40:06.703722275 +0000 UTC m=+5705.603509954" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.106677 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.535369 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.536784 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.539080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.539174 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.581645 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655385 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655498 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.655533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.704187 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.705601 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.711618 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.736752 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.745583 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.747079 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.756667 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757280 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757385 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757424 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757445 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757469 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757498 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.757517 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.763968 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.764162 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.764580 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.789399 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"nova-cell0-cell-mapping-lvbcr\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.811561 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.817333 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.819678 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.822467 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859151 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859199 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859325 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859396 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859424 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859449 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859467 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.859485 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.861003 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.862678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.862821 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.873923 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.882063 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.897950 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"nova-api-0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " pod="openstack/nova-api-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.950312 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.952020 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960778 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960809 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960835 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960863 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.960968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.961052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.964564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.965189 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.970959 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.971762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.971801 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.973028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.977169 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.981026 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.983785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"nova-metadata-0\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " pod="openstack/nova-metadata-0" Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.985008 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:40:11 crc kubenswrapper[4907]: I0313 15:40:11.987126 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"nova-scheduler-0\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.001175 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.006721 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.025475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.062858 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063193 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063219 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063248 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063269 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063300 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063320 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.063381 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.161381 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165194 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165401 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165470 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165528 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.165560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.167021 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.167652 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.169053 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.169310 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.175927 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.185925 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.186144 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"nova-cell1-novncproxy-0\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.186196 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"dnsmasq-dns-68dbfc785f-kjm62\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.319039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.329255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.491828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.599532 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.600562 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.602652 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.602842 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.630249 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:40:12 crc kubenswrapper[4907]: W0313 15:40:12.636377 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod188c81b4_cc35_444d_a5f1_f1633bae67a0.slice/crio-d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9 WatchSource:0}: Error finding container d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9: Status 404 returned error can't find the container with id d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9 Mar 13 15:40:12 crc kubenswrapper[4907]: W0313 15:40:12.640043 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5192153_c160_40bc_9375_a730e800e76e.slice/crio-a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242 WatchSource:0}: Error finding container a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242: Status 404 returned error can't find the container with id a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242 Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.645483 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.662661 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.677907 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.677968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.678291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.678413 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.725470 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.733357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerStarted","Data":"d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9"} Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.737254 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerStarted","Data":"a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242"} Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.738854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerStarted","Data":"2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc"} Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.738920 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerStarted","Data":"59f93412cb46a9a06d3f2d6b59dd6e308e6731badd1d9e181954832e856363ec"} Mar 13 15:40:12 crc kubenswrapper[4907]: W0313 15:40:12.744181 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc125541_0216_45f4_8276_520470f141fb.slice/crio-88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f WatchSource:0}: Error finding container 88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f: Status 404 returned error can't find the container with id 88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.762220 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lvbcr" podStartSLOduration=1.762199324 podStartE2EDuration="1.762199324s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:12.755658836 +0000 UTC m=+5711.655446525" watchObservedRunningTime="2026-03-13 15:40:12.762199324 +0000 UTC m=+5711.661987013" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780636 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780717 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780794 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.780838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.787210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.787818 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.788216 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.797229 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"nova-cell1-conductor-db-sync-c2mhb\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.869974 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.881588 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:40:12 crc kubenswrapper[4907]: I0313 15:40:12.938434 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:13 crc kubenswrapper[4907]: W0313 15:40:13.454507 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode935140e_65a6_48ff_9dda_fdf57edb9a4d.slice/crio-3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b WatchSource:0}: Error finding container 3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b: Status 404 returned error can't find the container with id 3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.455134 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.751284 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerStarted","Data":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.751332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerStarted","Data":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.753007 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerStarted","Data":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.753057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerStarted","Data":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.753074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerStarted","Data":"88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.754169 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerStarted","Data":"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.755352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerStarted","Data":"04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.755376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerStarted","Data":"3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.757770 4907 generic.go:334] "Generic (PLEG): container finished" podID="690a67d3-7115-4897-ba26-edd5eeff0521" containerID="5f757938e0484b76f29e92050d686900e13810e4b83c689080ec66591a9b77ed" exitCode=0 Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.757862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerDied","Data":"5f757938e0484b76f29e92050d686900e13810e4b83c689080ec66591a9b77ed"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.757896 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerStarted","Data":"4598382ec750564a06ecd379f171c64e71e4ac2c6f182c30bec9ff89463c8af3"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.765715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerStarted","Data":"2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.765755 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerStarted","Data":"97ff1fdcfd5fd91a658ff67b990106cf5c21eac69b149055f325885b224d558e"} Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.800013 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.799997609 podStartE2EDuration="2.799997609s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.799349581 +0000 UTC m=+5712.699137270" watchObservedRunningTime="2026-03-13 15:40:13.799997609 +0000 UTC m=+5712.699785298" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.857134 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.857117359 podStartE2EDuration="2.857117359s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.845022589 +0000 UTC m=+5712.744810278" watchObservedRunningTime="2026-03-13 15:40:13.857117359 +0000 UTC m=+5712.756905048" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.874807 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.874790152 podStartE2EDuration="2.874790152s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.871799721 +0000 UTC m=+5712.771587410" watchObservedRunningTime="2026-03-13 15:40:13.874790152 +0000 UTC m=+5712.774577841" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.925480 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" podStartSLOduration=1.925455747 podStartE2EDuration="1.925455747s" podCreationTimestamp="2026-03-13 15:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.921385865 +0000 UTC m=+5712.821173564" watchObservedRunningTime="2026-03-13 15:40:13.925455747 +0000 UTC m=+5712.825243436" Mar 13 15:40:13 crc kubenswrapper[4907]: I0313 15:40:13.949488 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.949468842 podStartE2EDuration="2.949468842s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:13.947327604 +0000 UTC m=+5712.847115313" watchObservedRunningTime="2026-03-13 15:40:13.949468842 +0000 UTC m=+5712.849256531" Mar 13 15:40:14 crc kubenswrapper[4907]: I0313 15:40:14.776229 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerStarted","Data":"b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce"} Mar 13 15:40:14 crc kubenswrapper[4907]: I0313 15:40:14.778053 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:14 crc kubenswrapper[4907]: I0313 15:40:14.802656 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" podStartSLOduration=3.8026372630000003 podStartE2EDuration="3.802637263s" podCreationTimestamp="2026-03-13 15:40:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:14.793543564 +0000 UTC m=+5713.693331253" watchObservedRunningTime="2026-03-13 15:40:14.802637263 +0000 UTC m=+5713.702424952" Mar 13 15:40:16 crc kubenswrapper[4907]: I0313 15:40:16.792537 4907 generic.go:334] "Generic (PLEG): container finished" podID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerID="04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e" exitCode=0 Mar 13 15:40:16 crc kubenswrapper[4907]: I0313 15:40:16.792687 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerDied","Data":"04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e"} Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.007753 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.329610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.802006 4907 generic.go:334] "Generic (PLEG): container finished" podID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerID="2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc" exitCode=0 Mar 13 15:40:17 crc kubenswrapper[4907]: I0313 15:40:17.802099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerDied","Data":"2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc"} Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.041714 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.042085 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.249133 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294577 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294651 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294692 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.294717 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") pod \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\" (UID: \"e935140e-65a6-48ff-9dda-fdf57edb9a4d\") " Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.304016 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts" (OuterVolumeSpecName: "scripts") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.304038 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm" (OuterVolumeSpecName: "kube-api-access-798cm") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "kube-api-access-798cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.321505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.339830 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data" (OuterVolumeSpecName: "config-data") pod "e935140e-65a6-48ff-9dda-fdf57edb9a4d" (UID: "e935140e-65a6-48ff-9dda-fdf57edb9a4d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.397978 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.398012 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.398027 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-798cm\" (UniqueName: \"kubernetes.io/projected/e935140e-65a6-48ff-9dda-fdf57edb9a4d-kube-api-access-798cm\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.398041 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e935140e-65a6-48ff-9dda-fdf57edb9a4d-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.813689 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" event={"ID":"e935140e-65a6-48ff-9dda-fdf57edb9a4d","Type":"ContainerDied","Data":"3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b"} Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.813745 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b30e6b6c78b39d41057f4a23478e828e3d6cb01d9da82b539e5864e0f6f7f4b" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.813700 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-c2mhb" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.922254 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:40:18 crc kubenswrapper[4907]: E0313 15:40:18.923336 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerName="nova-cell1-conductor-db-sync" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.923356 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerName="nova-cell1-conductor-db-sync" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.923611 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" containerName="nova-cell1-conductor-db-sync" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.925444 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.928042 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 15:40:18 crc kubenswrapper[4907]: I0313 15:40:18.935102 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.009339 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.009392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.009525 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.111032 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.111275 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.111418 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.116269 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.116275 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.127293 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"nova-cell1-conductor-0\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.185698 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.213361 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.213845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.213969 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.214173 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") pod \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\" (UID: \"89840ecf-ddb9-4a66-855b-0a1ae524bd6e\") " Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.217560 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts" (OuterVolumeSpecName: "scripts") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.221050 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk" (OuterVolumeSpecName: "kube-api-access-lvkgk") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "kube-api-access-lvkgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.241369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data" (OuterVolumeSpecName: "config-data") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.241438 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89840ecf-ddb9-4a66-855b-0a1ae524bd6e" (UID: "89840ecf-ddb9-4a66-855b-0a1ae524bd6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.248840 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316610 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316644 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvkgk\" (UniqueName: \"kubernetes.io/projected/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-kube-api-access-lvkgk\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316657 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.316665 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89840ecf-ddb9-4a66-855b-0a1ae524bd6e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.684786 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.823893 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerStarted","Data":"cbddd5aedb9a61c15acca2a6a299e779028d30303050e66d3356b0d2ec6abc1d"} Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.828763 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lvbcr" event={"ID":"89840ecf-ddb9-4a66-855b-0a1ae524bd6e","Type":"ContainerDied","Data":"59f93412cb46a9a06d3f2d6b59dd6e308e6731badd1d9e181954832e856363ec"} Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.828804 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59f93412cb46a9a06d3f2d6b59dd6e308e6731badd1d9e181954832e856363ec" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.828866 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lvbcr" Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.999235 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:19 crc kubenswrapper[4907]: I0313 15:40:19.999784 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" containerID="cri-o://08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.000003 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" containerID="cri-o://74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.008577 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.008776 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" containerID="cri-o://9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.065512 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.065733 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" containerID="cri-o://47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.065800 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" containerID="cri-o://1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" gracePeriod=30 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.529937 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.621524 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646131 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646290 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.646308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") pod \"188c81b4-cc35-444d-a5f1-f1633bae67a0\" (UID: \"188c81b4-cc35-444d-a5f1-f1633bae67a0\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.647159 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs" (OuterVolumeSpecName: "logs") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.651369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv" (OuterVolumeSpecName: "kube-api-access-6dmkv") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "kube-api-access-6dmkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.679104 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.689245 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data" (OuterVolumeSpecName: "config-data") pod "188c81b4-cc35-444d-a5f1-f1633bae67a0" (UID: "188c81b4-cc35-444d-a5f1-f1633bae67a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.747912 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.748323 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.748531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.748775 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") pod \"bc125541-0216-45f4-8276-520470f141fb\" (UID: \"bc125541-0216-45f4-8276-520470f141fb\") " Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.749299 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs" (OuterVolumeSpecName: "logs") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.749779 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc125541-0216-45f4-8276-520470f141fb-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.749936 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.750078 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188c81b4-cc35-444d-a5f1-f1633bae67a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.750170 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dmkv\" (UniqueName: \"kubernetes.io/projected/188c81b4-cc35-444d-a5f1-f1633bae67a0-kube-api-access-6dmkv\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.750251 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188c81b4-cc35-444d-a5f1-f1633bae67a0-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.753020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2" (OuterVolumeSpecName: "kube-api-access-hwqn2") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "kube-api-access-hwqn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.771809 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data" (OuterVolumeSpecName: "config-data") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.776163 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc125541-0216-45f4-8276-520470f141fb" (UID: "bc125541-0216-45f4-8276-520470f141fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.840705 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerStarted","Data":"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.841566 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.843873 4907 generic.go:334] "Generic (PLEG): container finished" podID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" exitCode=0 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.843933 4907 generic.go:334] "Generic (PLEG): container finished" podID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" exitCode=143 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.843943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerDied","Data":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.844062 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerDied","Data":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.844086 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188c81b4-cc35-444d-a5f1-f1633bae67a0","Type":"ContainerDied","Data":"d9e105ef07cb22db95da3ee8ed6caefc5f9cc867634b8baf1d1dfb328f44cec9"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.844145 4907 scope.go:117] "RemoveContainer" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846350 4907 generic.go:334] "Generic (PLEG): container finished" podID="bc125541-0216-45f4-8276-520470f141fb" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" exitCode=0 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846372 4907 generic.go:334] "Generic (PLEG): container finished" podID="bc125541-0216-45f4-8276-520470f141fb" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" exitCode=143 Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerDied","Data":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846415 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerDied","Data":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846425 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc125541-0216-45f4-8276-520470f141fb","Type":"ContainerDied","Data":"88a55c77e8165ad6fc6f2a546ffecd9fd6b597852354859b4d4ea8640528506f"} Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.846468 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.851067 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.852562 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwqn2\" (UniqueName: \"kubernetes.io/projected/bc125541-0216-45f4-8276-520470f141fb-kube-api-access-hwqn2\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.852587 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.852596 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc125541-0216-45f4-8276-520470f141fb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.861510 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.861491661 podStartE2EDuration="2.861491661s" podCreationTimestamp="2026-03-13 15:40:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:20.858172451 +0000 UTC m=+5719.757960140" watchObservedRunningTime="2026-03-13 15:40:20.861491661 +0000 UTC m=+5719.761279350" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.881718 4907 scope.go:117] "RemoveContainer" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.899482 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.914504 4907 scope.go:117] "RemoveContainer" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.915494 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": container with ID starting with 74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4 not found: ID does not exist" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.915568 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} err="failed to get container status \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": rpc error: code = NotFound desc = could not find container \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": container with ID starting with 74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.915627 4907 scope.go:117] "RemoveContainer" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.916270 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": container with ID starting with 08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b not found: ID does not exist" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.916335 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} err="failed to get container status \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": rpc error: code = NotFound desc = could not find container \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": container with ID starting with 08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.916353 4907 scope.go:117] "RemoveContainer" containerID="74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917253 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4"} err="failed to get container status \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": rpc error: code = NotFound desc = could not find container \"74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4\": container with ID starting with 74fbc4900d4783f6ccc5e6fc3b37fb82375ab950f69f896b8c992183e571e4e4 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917280 4907 scope.go:117] "RemoveContainer" containerID="08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917492 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b"} err="failed to get container status \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": rpc error: code = NotFound desc = could not find container \"08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b\": container with ID starting with 08fb2b974765a711e6b2fa7513d50a1469142e3763a394f769251a01b141f79b not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.917513 4907 scope.go:117] "RemoveContainer" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.920034 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.932787 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933311 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933335 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933351 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933359 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933372 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerName="nova-manage" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933379 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerName="nova-manage" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933396 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933404 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.933433 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933441 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933647 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933664 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" containerName="nova-api-api" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933678 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-metadata" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933696 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" containerName="nova-manage" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.933720 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc125541-0216-45f4-8276-520470f141fb" containerName="nova-metadata-log" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.935026 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.942610 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.944248 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.949541 4907 scope.go:117] "RemoveContainer" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.960604 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.980242 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.984777 4907 scope.go:117] "RemoveContainer" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.985276 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": container with ID starting with 1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373 not found: ID does not exist" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985310 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} err="failed to get container status \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": rpc error: code = NotFound desc = could not find container \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": container with ID starting with 1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985333 4907 scope.go:117] "RemoveContainer" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: E0313 15:40:20.985725 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": container with ID starting with 47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba not found: ID does not exist" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985745 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} err="failed to get container status \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": rpc error: code = NotFound desc = could not find container \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": container with ID starting with 47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.985759 4907 scope.go:117] "RemoveContainer" containerID="1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.986109 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373"} err="failed to get container status \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": rpc error: code = NotFound desc = could not find container \"1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373\": container with ID starting with 1ca5f3e9f3e3e7e3febfd017c86076ba77be5580f4a8b79cbb7bc8952ab76373 not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.986138 4907 scope.go:117] "RemoveContainer" containerID="47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.986451 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba"} err="failed to get container status \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": rpc error: code = NotFound desc = could not find container \"47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba\": container with ID starting with 47b10daf3ce966a6fd75bc5149bce0256a8ca53a79611f00cde77948ce92e7ba not found: ID does not exist" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.994131 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.995738 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.997907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:40:20 crc kubenswrapper[4907]: I0313 15:40:20.997976 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055330 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055428 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055617 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055715 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.055840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158140 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158203 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158221 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158583 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158603 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.158791 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.159204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.162743 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.163349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.164237 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.174660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.175097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"nova-metadata-0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.178926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"nova-api-0\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.271107 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.319956 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.718862 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.797239 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188c81b4-cc35-444d-a5f1-f1633bae67a0" path="/var/lib/kubelet/pods/188c81b4-cc35-444d-a5f1-f1633bae67a0/volumes" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.798804 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc125541-0216-45f4-8276-520470f141fb" path="/var/lib/kubelet/pods/bc125541-0216-45f4-8276-520470f141fb/volumes" Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.799946 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.854586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerStarted","Data":"0d9dea632a506015513087187b39d9560b7d9208b5328332f68e9033c985b609"} Mar 13 15:40:21 crc kubenswrapper[4907]: I0313 15:40:21.858316 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerStarted","Data":"7d15b64a01f946210967bf64f6d3e77361db086303536eb844ce8563c1a6da6d"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.320086 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.330483 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.346176 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.399778 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.400954 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" containerID="cri-o://4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" gracePeriod=10 Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.874111 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.881222 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerStarted","Data":"e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.881259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerStarted","Data":"a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.884620 4907 generic.go:334] "Generic (PLEG): container finished" podID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" exitCode=0 Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.884679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerDied","Data":"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.884849 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.885083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c58f95-579qq" event={"ID":"5153eb33-98f7-4452-b4f5-751a6d49c54b","Type":"ContainerDied","Data":"1f607dd9ca9be04a0a9e3a628ae846054d952f5ec49fd5e3e8f6f1422eb9fe8b"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.885106 4907 scope.go:117] "RemoveContainer" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.889059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerStarted","Data":"9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.889083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerStarted","Data":"12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c"} Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.901790 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.938010 4907 scope.go:117] "RemoveContainer" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.948447 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.94842752 podStartE2EDuration="2.94842752s" podCreationTimestamp="2026-03-13 15:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:22.924119085 +0000 UTC m=+5721.823906794" watchObservedRunningTime="2026-03-13 15:40:22.94842752 +0000 UTC m=+5721.848215209" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.950239 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.950231039 podStartE2EDuration="2.950231039s" podCreationTimestamp="2026-03-13 15:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:22.947371561 +0000 UTC m=+5721.847159250" watchObservedRunningTime="2026-03-13 15:40:22.950231039 +0000 UTC m=+5721.850018728" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.968983 4907 scope.go:117] "RemoveContainer" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" Mar 13 15:40:22 crc kubenswrapper[4907]: E0313 15:40:22.975512 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59\": container with ID starting with 4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59 not found: ID does not exist" containerID="4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.975565 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59"} err="failed to get container status \"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59\": rpc error: code = NotFound desc = could not find container \"4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59\": container with ID starting with 4d47e9b4f5705f84e359bebffdaf6b40c2a5a7d780d8daa8048b0a559dabed59 not found: ID does not exist" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.975597 4907 scope.go:117] "RemoveContainer" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" Mar 13 15:40:22 crc kubenswrapper[4907]: E0313 15:40:22.975992 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78\": container with ID starting with dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78 not found: ID does not exist" containerID="dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.976019 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78"} err="failed to get container status \"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78\": rpc error: code = NotFound desc = could not find container \"dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78\": container with ID starting with dc68df440400c3300857779e0180b105a0536c85adecc574069e0c384c4a2b78 not found: ID does not exist" Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.989755 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.989832 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.989968 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.990006 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:22 crc kubenswrapper[4907]: I0313 15:40:22.990032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") pod \"5153eb33-98f7-4452-b4f5-751a6d49c54b\" (UID: \"5153eb33-98f7-4452-b4f5-751a6d49c54b\") " Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.009008 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr" (OuterVolumeSpecName: "kube-api-access-gskhr") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "kube-api-access-gskhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.052157 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.053963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config" (OuterVolumeSpecName: "config") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.061630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.064667 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5153eb33-98f7-4452-b4f5-751a6d49c54b" (UID: "5153eb33-98f7-4452-b4f5-751a6d49c54b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091683 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091710 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gskhr\" (UniqueName: \"kubernetes.io/projected/5153eb33-98f7-4452-b4f5-751a6d49c54b-kube-api-access-gskhr\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091722 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091731 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.091741 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5153eb33-98f7-4452-b4f5-751a6d49c54b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.227405 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.235281 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dd7c58f95-579qq"] Mar 13 15:40:23 crc kubenswrapper[4907]: I0313 15:40:23.792390 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" path="/var/lib/kubelet/pods/5153eb33-98f7-4452-b4f5-751a6d49c54b/volumes" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.277578 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.698839 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.767414 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.767859 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="init" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.767958 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="init" Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.767988 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.767999 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.768018 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.768027 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.768280 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5153eb33-98f7-4452-b4f5-751a6d49c54b" containerName="dnsmasq-dns" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.768309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5192153-c160-40bc-9375-a730e800e76e" containerName="nova-scheduler-scheduler" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.769042 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.777332 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.777560 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.780715 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.824777 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") pod \"e5192153-c160-40bc-9375-a730e800e76e\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.824861 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") pod \"e5192153-c160-40bc-9375-a730e800e76e\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.824977 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") pod \"e5192153-c160-40bc-9375-a730e800e76e\" (UID: \"e5192153-c160-40bc-9375-a730e800e76e\") " Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.838786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j" (OuterVolumeSpecName: "kube-api-access-5kk9j") pod "e5192153-c160-40bc-9375-a730e800e76e" (UID: "e5192153-c160-40bc-9375-a730e800e76e"). InnerVolumeSpecName "kube-api-access-5kk9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.851754 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5192153-c160-40bc-9375-a730e800e76e" (UID: "e5192153-c160-40bc-9375-a730e800e76e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.852137 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data" (OuterVolumeSpecName: "config-data") pod "e5192153-c160-40bc-9375-a730e800e76e" (UID: "e5192153-c160-40bc-9375-a730e800e76e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910596 4907 generic.go:334] "Generic (PLEG): container finished" podID="e5192153-c160-40bc-9375-a730e800e76e" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" exitCode=0 Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910643 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerDied","Data":"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca"} Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910673 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5192153-c160-40bc-9375-a730e800e76e","Type":"ContainerDied","Data":"a6f0a44a2f3b94386ebd8b9c603db045564e6d31764a5593228f18f153118242"} Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910690 4907 scope.go:117] "RemoveContainer" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.910817 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.929019 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.929253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.930048 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932706 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932895 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932919 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kk9j\" (UniqueName: \"kubernetes.io/projected/e5192153-c160-40bc-9375-a730e800e76e-kube-api-access-5kk9j\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.932933 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5192153-c160-40bc-9375-a730e800e76e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.948559 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.954525 4907 scope.go:117] "RemoveContainer" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" Mar 13 15:40:24 crc kubenswrapper[4907]: E0313 15:40:24.955146 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca\": container with ID starting with 9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca not found: ID does not exist" containerID="9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.955192 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca"} err="failed to get container status \"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca\": rpc error: code = NotFound desc = could not find container \"9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca\": container with ID starting with 9ca8fe863388471bc1a4b09ca8e4b518970b726d2835186fd66e2f90d2cda6ca not found: ID does not exist" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.967608 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.977243 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.978596 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:24 crc kubenswrapper[4907]: I0313 15:40:24.981150 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.002629 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033323 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033367 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033418 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033444 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033560 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033666 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.033726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.037262 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.037292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.037454 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.047836 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"nova-cell1-cell-mapping-d9b97\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.095903 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.137012 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.137154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.137379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.158495 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.158619 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.159346 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.310761 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.566287 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.823449 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5192153-c160-40bc-9375-a730e800e76e" path="/var/lib/kubelet/pods/e5192153-c160-40bc-9375-a730e800e76e/volumes" Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.860998 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:25 crc kubenswrapper[4907]: W0313 15:40:25.866961 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b2e0902_2ec8_4712_8215_71643998f9ae.slice/crio-f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93 WatchSource:0}: Error finding container f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93: Status 404 returned error can't find the container with id f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93 Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.923387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerStarted","Data":"d3ac5726910e7451a7d9a4ee3678b7f325b341f8f577f2d9ded3dcb9a624377a"} Mar 13 15:40:25 crc kubenswrapper[4907]: I0313 15:40:25.925325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerStarted","Data":"f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93"} Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.945337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerStarted","Data":"0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660"} Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.948498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerStarted","Data":"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32"} Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.971592 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-d9b97" podStartSLOduration=2.971573059 podStartE2EDuration="2.971573059s" podCreationTimestamp="2026-03-13 15:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:26.966520331 +0000 UTC m=+5725.866308040" watchObservedRunningTime="2026-03-13 15:40:26.971573059 +0000 UTC m=+5725.871360738" Mar 13 15:40:26 crc kubenswrapper[4907]: I0313 15:40:26.988331 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.988307336 podStartE2EDuration="2.988307336s" podCreationTimestamp="2026-03-13 15:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:26.983132435 +0000 UTC m=+5725.882920134" watchObservedRunningTime="2026-03-13 15:40:26.988307336 +0000 UTC m=+5725.888095025" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.660444 4907 scope.go:117] "RemoveContainer" containerID="a3109c92b6a4f47975f53b04558486733a8cd9a2cb768f638be94b37e0295965" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.684245 4907 scope.go:117] "RemoveContainer" containerID="47fa5a52c38336d054a9aed448aae48b3dd0ee12065b60aaad1abb03635d9d32" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.750167 4907 scope.go:117] "RemoveContainer" containerID="7829d0510750d5cae8df3422b54b042bd33e40f132a8e3e16f6eec0a8952963b" Mar 13 15:40:28 crc kubenswrapper[4907]: I0313 15:40:28.781547 4907 scope.go:117] "RemoveContainer" containerID="1eda03ad73ea12900c383cf8f9715d8d1c4690e9fdc026cd78850ed84ad83b37" Mar 13 15:40:30 crc kubenswrapper[4907]: I0313 15:40:30.311855 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:40:30 crc kubenswrapper[4907]: I0313 15:40:30.995711 4907 generic.go:334] "Generic (PLEG): container finished" podID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerID="0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660" exitCode=0 Mar 13 15:40:30 crc kubenswrapper[4907]: I0313 15:40:30.995804 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerDied","Data":"0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660"} Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.272317 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.273598 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.320539 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:31 crc kubenswrapper[4907]: I0313 15:40:31.320593 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.355085 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.108:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.355119 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.108:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.360704 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.438165 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.109:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.438766 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.109:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480246 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480305 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480373 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.480397 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") pod \"f6e87a99-1744-4002-a08b-dabf8a97e48b\" (UID: \"f6e87a99-1744-4002-a08b-dabf8a97e48b\") " Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.486998 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts" (OuterVolumeSpecName: "scripts") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.487131 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv" (OuterVolumeSpecName: "kube-api-access-vl2kv") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "kube-api-access-vl2kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.517152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data" (OuterVolumeSpecName: "config-data") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.525747 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6e87a99-1744-4002-a08b-dabf8a97e48b" (UID: "f6e87a99-1744-4002-a08b-dabf8a97e48b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581836 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581870 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl2kv\" (UniqueName: \"kubernetes.io/projected/f6e87a99-1744-4002-a08b-dabf8a97e48b-kube-api-access-vl2kv\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581898 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:32 crc kubenswrapper[4907]: I0313 15:40:32.581907 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6e87a99-1744-4002-a08b-dabf8a97e48b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.014727 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d9b97" event={"ID":"f6e87a99-1744-4002-a08b-dabf8a97e48b","Type":"ContainerDied","Data":"d3ac5726910e7451a7d9a4ee3678b7f325b341f8f577f2d9ded3dcb9a624377a"} Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.014769 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3ac5726910e7451a7d9a4ee3678b7f325b341f8f577f2d9ded3dcb9a624377a" Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.014775 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d9b97" Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.209929 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.210434 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" containerID="cri-o://a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.210500 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" containerID="cri-o://e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.256164 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.256392 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" containerID="cri-o://1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.272259 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.272550 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" containerID="cri-o://12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c" gracePeriod=30 Mar 13 15:40:33 crc kubenswrapper[4907]: I0313 15:40:33.272739 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" containerID="cri-o://9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d" gracePeriod=30 Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.025451 4907 generic.go:334] "Generic (PLEG): container finished" podID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerID="a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147" exitCode=143 Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.025531 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerDied","Data":"a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147"} Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.027452 4907 generic.go:334] "Generic (PLEG): container finished" podID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerID="12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c" exitCode=143 Mar 13 15:40:34 crc kubenswrapper[4907]: I0313 15:40:34.027490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerDied","Data":"12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c"} Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.636732 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.782854 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") pod \"6b2e0902-2ec8-4712-8215-71643998f9ae\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.782967 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") pod \"6b2e0902-2ec8-4712-8215-71643998f9ae\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.783161 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") pod \"6b2e0902-2ec8-4712-8215-71643998f9ae\" (UID: \"6b2e0902-2ec8-4712-8215-71643998f9ae\") " Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.788473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl" (OuterVolumeSpecName: "kube-api-access-lpzzl") pod "6b2e0902-2ec8-4712-8215-71643998f9ae" (UID: "6b2e0902-2ec8-4712-8215-71643998f9ae"). InnerVolumeSpecName "kube-api-access-lpzzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.809130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data" (OuterVolumeSpecName: "config-data") pod "6b2e0902-2ec8-4712-8215-71643998f9ae" (UID: "6b2e0902-2ec8-4712-8215-71643998f9ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.810926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b2e0902-2ec8-4712-8215-71643998f9ae" (UID: "6b2e0902-2ec8-4712-8215-71643998f9ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.885676 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.885699 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b2e0902-2ec8-4712-8215-71643998f9ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:37 crc kubenswrapper[4907]: I0313 15:40:37.885709 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpzzl\" (UniqueName: \"kubernetes.io/projected/6b2e0902-2ec8-4712-8215-71643998f9ae-kube-api-access-lpzzl\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.063931 4907 generic.go:334] "Generic (PLEG): container finished" podID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" exitCode=0 Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064014 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064014 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerDied","Data":"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b2e0902-2ec8-4712-8215-71643998f9ae","Type":"ContainerDied","Data":"f8adfb940d43d79ae7a52f858d549a2c2afbabbdf09ea298e87ae615f8153c93"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.064200 4907 scope.go:117] "RemoveContainer" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.067571 4907 generic.go:334] "Generic (PLEG): container finished" podID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerID="e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df" exitCode=0 Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.067626 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerDied","Data":"e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.071467 4907 generic.go:334] "Generic (PLEG): container finished" podID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerID="9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d" exitCode=0 Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.071626 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerDied","Data":"9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d"} Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.088231 4907 scope.go:117] "RemoveContainer" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.093839 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32\": container with ID starting with 1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32 not found: ID does not exist" containerID="1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.093947 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32"} err="failed to get container status \"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32\": rpc error: code = NotFound desc = could not find container \"1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32\": container with ID starting with 1f04df211f07bf45e057e613b89c2ab160d4ab249aeb981eda1a7e92b8868b32 not found: ID does not exist" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.105191 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.126660 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.128453 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.138938 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139345 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139359 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139371 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139390 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139396 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" Mar 13 15:40:38 crc kubenswrapper[4907]: E0313 15:40:38.139422 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerName="nova-manage" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139428 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerName="nova-manage" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139581 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-metadata" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139591 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" containerName="nova-manage" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139607 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" containerName="nova-scheduler-scheduler" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.139622 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" containerName="nova-metadata-log" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.140263 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.142330 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.148555 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214273 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214338 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214542 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") pod \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\" (UID: \"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214733 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214871 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.214914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.215608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs" (OuterVolumeSpecName: "logs") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.221917 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk" (OuterVolumeSpecName: "kube-api-access-nm2jk") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "kube-api-access-nm2jk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.242748 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.253138 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.253844 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data" (OuterVolumeSpecName: "config-data") pod "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" (UID: "8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316055 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316234 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm2jk\" (UniqueName: \"kubernetes.io/projected/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-kube-api-access-nm2jk\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316245 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316255 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.316263 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.319684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.323807 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.331336 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"nova-scheduler-0\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417260 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417352 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417381 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.417482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") pod \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\" (UID: \"ad3c8be4-d237-4c1e-bd98-963acae4ac32\") " Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.418011 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs" (OuterVolumeSpecName: "logs") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.421011 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv" (OuterVolumeSpecName: "kube-api-access-x6msv") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "kube-api-access-x6msv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.439492 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.439555 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data" (OuterVolumeSpecName: "config-data") pod "ad3c8be4-d237-4c1e-bd98-963acae4ac32" (UID: "ad3c8be4-d237-4c1e-bd98-963acae4ac32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519495 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519541 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad3c8be4-d237-4c1e-bd98-963acae4ac32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519558 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ad3c8be4-d237-4c1e-bd98-963acae4ac32-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.519570 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6msv\" (UniqueName: \"kubernetes.io/projected/ad3c8be4-d237-4c1e-bd98-963acae4ac32-kube-api-access-x6msv\") on node \"crc\" DevicePath \"\"" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.520570 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:40:38 crc kubenswrapper[4907]: I0313 15:40:38.957327 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:40:38 crc kubenswrapper[4907]: W0313 15:40:38.960555 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61a2ddd9_1fcc_427e_aaa2_f092ebb7bd30.slice/crio-605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725 WatchSource:0}: Error finding container 605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725: Status 404 returned error can't find the container with id 605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725 Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.082051 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerStarted","Data":"605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725"} Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.084441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ad3c8be4-d237-4c1e-bd98-963acae4ac32","Type":"ContainerDied","Data":"7d15b64a01f946210967bf64f6d3e77361db086303536eb844ce8563c1a6da6d"} Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.084484 4907 scope.go:117] "RemoveContainer" containerID="e833237c58d59cd7a382ff7644c38c3979cbe4fa6a8a9e044f9040c86f3823df" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.084522 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.090987 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0","Type":"ContainerDied","Data":"0d9dea632a506015513087187b39d9560b7d9208b5328332f68e9033c985b609"} Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.091130 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.123092 4907 scope.go:117] "RemoveContainer" containerID="a9b52fcfd1633f1cfd4fe4e130d7daf6a6a0f261d982b7490f4c49360b713147" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.144976 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.162163 4907 scope.go:117] "RemoveContainer" containerID="9d4f1d5de2313cc0cacaa0de90599bf7cc29209144c55b94e7e5f63f34cd569d" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.173214 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.185359 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.201533 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: E0313 15:40:39.201973 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.201992 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" Mar 13 15:40:39 crc kubenswrapper[4907]: E0313 15:40:39.202031 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.202038 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.202210 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-log" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.202232 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" containerName="nova-api-api" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.203227 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.203963 4907 scope.go:117] "RemoveContainer" containerID="12ad1f3ba8429b32fe4517953f072fa40cf2ef694f54ab5cea4b98ef8339414c" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.212074 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.215105 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.228388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232663 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232766 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.232923 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.236705 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.238615 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.242718 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.246483 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335053 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335111 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335156 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335315 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335473 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.335902 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.338427 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.339758 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.340452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.351321 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"nova-api-0\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439700 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439823 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.439949 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.440427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.443474 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.452251 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.455337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"nova-metadata-0\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.525150 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.558441 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.814615 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b2e0902-2ec8-4712-8215-71643998f9ae" path="/var/lib/kubelet/pods/6b2e0902-2ec8-4712-8215-71643998f9ae/volumes" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.818522 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0" path="/var/lib/kubelet/pods/8afb3ee7-4e5e-4f1c-a703-d082d9fbccd0/volumes" Mar 13 15:40:39 crc kubenswrapper[4907]: I0313 15:40:39.821033 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad3c8be4-d237-4c1e-bd98-963acae4ac32" path="/var/lib/kubelet/pods/ad3c8be4-d237-4c1e-bd98-963acae4ac32/volumes" Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.008110 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:40:40 crc kubenswrapper[4907]: W0313 15:40:40.009630 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81ef3cd7_e06f_49f8_bc01_41b32fbb5546.slice/crio-c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9 WatchSource:0}: Error finding container c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9: Status 404 returned error can't find the container with id c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9 Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.016148 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:40:40 crc kubenswrapper[4907]: W0313 15:40:40.023958 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37198210_23ed_4025_8cd9_e284c6573318.slice/crio-54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5 WatchSource:0}: Error finding container 54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5: Status 404 returned error can't find the container with id 54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5 Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.116429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerStarted","Data":"54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5"} Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.119482 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerStarted","Data":"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469"} Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.125018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerStarted","Data":"c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9"} Mar 13 15:40:40 crc kubenswrapper[4907]: I0313 15:40:40.146788 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.146746347 podStartE2EDuration="2.146746347s" podCreationTimestamp="2026-03-13 15:40:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:40.135067898 +0000 UTC m=+5739.034855577" watchObservedRunningTime="2026-03-13 15:40:40.146746347 +0000 UTC m=+5739.046534036" Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.134174 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerStarted","Data":"ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.134502 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerStarted","Data":"6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.136432 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerStarted","Data":"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.136476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerStarted","Data":"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358"} Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.155936 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.155911819 podStartE2EDuration="2.155911819s" podCreationTimestamp="2026-03-13 15:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:41.149894134 +0000 UTC m=+5740.049681823" watchObservedRunningTime="2026-03-13 15:40:41.155911819 +0000 UTC m=+5740.055699508" Mar 13 15:40:41 crc kubenswrapper[4907]: I0313 15:40:41.196659 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.196630962 podStartE2EDuration="2.196630962s" podCreationTimestamp="2026-03-13 15:40:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:40:41.176211704 +0000 UTC m=+5740.075999393" watchObservedRunningTime="2026-03-13 15:40:41.196630962 +0000 UTC m=+5740.096418651" Mar 13 15:40:43 crc kubenswrapper[4907]: I0313 15:40:43.521583 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.041374 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.043349 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.521301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 15:40:48 crc kubenswrapper[4907]: I0313 15:40:48.551369 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.235972 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.526079 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.526426 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.559421 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:49 crc kubenswrapper[4907]: I0313 15:40:49.559617 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.609166 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.113:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.692133 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.114:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.692315 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.114:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:50 crc kubenswrapper[4907]: I0313 15:40:50.692133 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.113:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.526203 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.526612 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.559842 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:40:57 crc kubenswrapper[4907]: I0313 15:40:57.560015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.529560 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.532141 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.534607 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.562091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.563480 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:40:59 crc kubenswrapper[4907]: I0313 15:40:59.569124 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.294931 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.295845 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.496466 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.499662 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.526545 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625696 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625775 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625809 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.625929 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727311 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.727349 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728379 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728407 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.728480 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.749102 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"dnsmasq-dns-f9699695f-pwxvd\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:00 crc kubenswrapper[4907]: I0313 15:41:00.831050 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:01 crc kubenswrapper[4907]: I0313 15:41:01.330451 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:01 crc kubenswrapper[4907]: W0313 15:41:01.334988 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda122794c_3b28_4344_9817_f05a9e663110.slice/crio-8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e WatchSource:0}: Error finding container 8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e: Status 404 returned error can't find the container with id 8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e Mar 13 15:41:02 crc kubenswrapper[4907]: I0313 15:41:02.312035 4907 generic.go:334] "Generic (PLEG): container finished" podID="a122794c-3b28-4344-9817-f05a9e663110" containerID="ec3db5cd0d09990bedd2c273757078f6b84cc7be565b5f045a0e95e3abe11b83" exitCode=0 Mar 13 15:41:02 crc kubenswrapper[4907]: I0313 15:41:02.312153 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerDied","Data":"ec3db5cd0d09990bedd2c273757078f6b84cc7be565b5f045a0e95e3abe11b83"} Mar 13 15:41:02 crc kubenswrapper[4907]: I0313 15:41:02.312601 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerStarted","Data":"8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e"} Mar 13 15:41:03 crc kubenswrapper[4907]: I0313 15:41:03.322564 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerStarted","Data":"73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd"} Mar 13 15:41:03 crc kubenswrapper[4907]: I0313 15:41:03.322922 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:03 crc kubenswrapper[4907]: I0313 15:41:03.340069 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" podStartSLOduration=3.340045119 podStartE2EDuration="3.340045119s" podCreationTimestamp="2026-03-13 15:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:03.337305744 +0000 UTC m=+5762.237093433" watchObservedRunningTime="2026-03-13 15:41:03.340045119 +0000 UTC m=+5762.239832848" Mar 13 15:41:10 crc kubenswrapper[4907]: I0313 15:41:10.833039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:10 crc kubenswrapper[4907]: I0313 15:41:10.893298 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:41:10 crc kubenswrapper[4907]: I0313 15:41:10.893529 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" containerID="cri-o://b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce" gracePeriod=10 Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.419807 4907 generic.go:334] "Generic (PLEG): container finished" podID="690a67d3-7115-4897-ba26-edd5eeff0521" containerID="b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce" exitCode=0 Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.420112 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerDied","Data":"b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce"} Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.549990 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.634813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.634932 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.634964 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.635036 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.635082 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") pod \"690a67d3-7115-4897-ba26-edd5eeff0521\" (UID: \"690a67d3-7115-4897-ba26-edd5eeff0521\") " Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.642636 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq" (OuterVolumeSpecName: "kube-api-access-x55kq") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "kube-api-access-x55kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.690801 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config" (OuterVolumeSpecName: "config") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.702813 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.708770 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.727571 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "690a67d3-7115-4897-ba26-edd5eeff0521" (UID: "690a67d3-7115-4897-ba26-edd5eeff0521"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742588 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742628 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742664 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742677 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55kq\" (UniqueName: \"kubernetes.io/projected/690a67d3-7115-4897-ba26-edd5eeff0521-kube-api-access-x55kq\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:11 crc kubenswrapper[4907]: I0313 15:41:11.742691 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/690a67d3-7115-4897-ba26-edd5eeff0521-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.429621 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" event={"ID":"690a67d3-7115-4897-ba26-edd5eeff0521","Type":"ContainerDied","Data":"4598382ec750564a06ecd379f171c64e71e4ac2c6f182c30bec9ff89463c8af3"} Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.429937 4907 scope.go:117] "RemoveContainer" containerID="b0ec28b6beaa68b816d49abaf6a1073e21a0e38ace3bf382a509d710008100ce" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.429673 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68dbfc785f-kjm62" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.456807 4907 scope.go:117] "RemoveContainer" containerID="5f757938e0484b76f29e92050d686900e13810e4b83c689080ec66591a9b77ed" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.485261 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.494785 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68dbfc785f-kjm62"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833001 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:41:12 crc kubenswrapper[4907]: E0313 15:41:12.833361 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="init" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833378 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="init" Mar 13 15:41:12 crc kubenswrapper[4907]: E0313 15:41:12.833411 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833417 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.833587 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" containerName="dnsmasq-dns" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.834160 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.850281 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.861157 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.861420 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.940294 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.941743 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.943770 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.959925 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.963429 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.963478 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.964261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:12 crc kubenswrapper[4907]: I0313 15:41:12.981253 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"cinder-db-create-pvpwm\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.065340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.065387 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.150178 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.167252 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.167423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.168214 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.184260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"cinder-9290-account-create-update-qws42\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.263376 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.600496 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.709761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:41:13 crc kubenswrapper[4907]: W0313 15:41:13.719411 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ab8c7fa_9565_4b6c_b8f8_033cfdf444db.slice/crio-80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881 WatchSource:0}: Error finding container 80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881: Status 404 returned error can't find the container with id 80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881 Mar 13 15:41:13 crc kubenswrapper[4907]: I0313 15:41:13.800676 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690a67d3-7115-4897-ba26-edd5eeff0521" path="/var/lib/kubelet/pods/690a67d3-7115-4897-ba26-edd5eeff0521/volumes" Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.471946 4907 generic.go:334] "Generic (PLEG): container finished" podID="499bc96f-8fa9-4900-bd92-2a507215791f" containerID="5140f9e51d2f4e22927da88ea13b8b76a7a069356e41a43262ca41c166a3779c" exitCode=0 Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.473165 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pvpwm" event={"ID":"499bc96f-8fa9-4900-bd92-2a507215791f","Type":"ContainerDied","Data":"5140f9e51d2f4e22927da88ea13b8b76a7a069356e41a43262ca41c166a3779c"} Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.473276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pvpwm" event={"ID":"499bc96f-8fa9-4900-bd92-2a507215791f","Type":"ContainerStarted","Data":"4f157b815232d30b0784ee5079591971c6c531c2992a38589627a781beae23d9"} Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.476256 4907 generic.go:334] "Generic (PLEG): container finished" podID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerID="0559fde0f7d2daa5d36de2042211ef456904ceea0804232b86a43826c351c91c" exitCode=0 Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.476265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9290-account-create-update-qws42" event={"ID":"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db","Type":"ContainerDied","Data":"0559fde0f7d2daa5d36de2042211ef456904ceea0804232b86a43826c351c91c"} Mar 13 15:41:14 crc kubenswrapper[4907]: I0313 15:41:14.476338 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9290-account-create-update-qws42" event={"ID":"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db","Type":"ContainerStarted","Data":"80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881"} Mar 13 15:41:15 crc kubenswrapper[4907]: I0313 15:41:15.956105 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:15 crc kubenswrapper[4907]: I0313 15:41:15.966020 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") pod \"499bc96f-8fa9-4900-bd92-2a507215791f\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019644 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") pod \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019673 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") pod \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\" (UID: \"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.019778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") pod \"499bc96f-8fa9-4900-bd92-2a507215791f\" (UID: \"499bc96f-8fa9-4900-bd92-2a507215791f\") " Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.021034 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "499bc96f-8fa9-4900-bd92-2a507215791f" (UID: "499bc96f-8fa9-4900-bd92-2a507215791f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.021130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" (UID: "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.027303 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz" (OuterVolumeSpecName: "kube-api-access-rvbzz") pod "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" (UID: "0ab8c7fa-9565-4b6c-b8f8-033cfdf444db"). InnerVolumeSpecName "kube-api-access-rvbzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.027659 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg" (OuterVolumeSpecName: "kube-api-access-52hkg") pod "499bc96f-8fa9-4900-bd92-2a507215791f" (UID: "499bc96f-8fa9-4900-bd92-2a507215791f"). InnerVolumeSpecName "kube-api-access-52hkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52hkg\" (UniqueName: \"kubernetes.io/projected/499bc96f-8fa9-4900-bd92-2a507215791f-kube-api-access-52hkg\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121698 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121707 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvbzz\" (UniqueName: \"kubernetes.io/projected/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db-kube-api-access-rvbzz\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.121715 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/499bc96f-8fa9-4900-bd92-2a507215791f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.496137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pvpwm" event={"ID":"499bc96f-8fa9-4900-bd92-2a507215791f","Type":"ContainerDied","Data":"4f157b815232d30b0784ee5079591971c6c531c2992a38589627a781beae23d9"} Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.496478 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f157b815232d30b0784ee5079591971c6c531c2992a38589627a781beae23d9" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.496191 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pvpwm" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.497718 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-9290-account-create-update-qws42" event={"ID":"0ab8c7fa-9565-4b6c-b8f8-033cfdf444db","Type":"ContainerDied","Data":"80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881"} Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.497746 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80540ba04d58c62659d9d51465d46822d741fb9ff2667a93002e21e7b5205881" Mar 13 15:41:16 crc kubenswrapper[4907]: I0313 15:41:16.497797 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-9290-account-create-update-qws42" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.041336 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.041402 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.041457 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.042265 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.042320 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" gracePeriod=600 Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.190757 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.197980 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.198777 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" containerName="mariadb-database-create" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.198807 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" containerName="mariadb-database-create" Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.198851 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerName="mariadb-account-create-update" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.198861 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerName="mariadb-account-create-update" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.201102 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" containerName="mariadb-database-create" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.201173 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" containerName="mariadb-account-create-update" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.204790 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.209148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v5jht" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.211000 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.213914 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.219607 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260555 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260657 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260787 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.260991 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.261199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363472 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363549 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363616 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363633 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363651 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.363710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.368501 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.368737 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.369214 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.375429 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.380506 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"cinder-db-sync-hv2mf\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.515021 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" exitCode=0 Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.515089 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5"} Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.515140 4907 scope.go:117] "RemoveContainer" containerID="90a39cf81ab6b6b036397a9c751a635e5f4cd834dccdc53b2ff02dd0144698a4" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.516178 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:18 crc kubenswrapper[4907]: E0313 15:41:18.516490 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:18 crc kubenswrapper[4907]: I0313 15:41:18.530870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:19 crc kubenswrapper[4907]: I0313 15:41:19.038356 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:41:19 crc kubenswrapper[4907]: I0313 15:41:19.523846 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerStarted","Data":"d1248fb6962eeda0ac2a1465b0501825fb69861c03d06ab1f3e3c6b244ab9af9"} Mar 13 15:41:20 crc kubenswrapper[4907]: I0313 15:41:20.533954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerStarted","Data":"cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877"} Mar 13 15:41:20 crc kubenswrapper[4907]: I0313 15:41:20.564741 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-hv2mf" podStartSLOduration=2.564716546 podStartE2EDuration="2.564716546s" podCreationTimestamp="2026-03-13 15:41:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:20.553754687 +0000 UTC m=+5779.453542376" watchObservedRunningTime="2026-03-13 15:41:20.564716546 +0000 UTC m=+5779.464504265" Mar 13 15:41:22 crc kubenswrapper[4907]: I0313 15:41:22.552053 4907 generic.go:334] "Generic (PLEG): container finished" podID="ef105bad-6763-4234-b52b-6d2820d48b02" containerID="cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877" exitCode=0 Mar 13 15:41:22 crc kubenswrapper[4907]: I0313 15:41:22.552374 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerDied","Data":"cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877"} Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.927413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974463 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974530 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974581 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974640 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.974667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") pod \"ef105bad-6763-4234-b52b-6d2820d48b02\" (UID: \"ef105bad-6763-4234-b52b-6d2820d48b02\") " Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.975026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.980290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.981794 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4" (OuterVolumeSpecName: "kube-api-access-skdq4") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "kube-api-access-skdq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:23 crc kubenswrapper[4907]: I0313 15:41:23.993436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts" (OuterVolumeSpecName: "scripts") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.006234 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.037229 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data" (OuterVolumeSpecName: "config-data") pod "ef105bad-6763-4234-b52b-6d2820d48b02" (UID: "ef105bad-6763-4234-b52b-6d2820d48b02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075854 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075909 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skdq4\" (UniqueName: \"kubernetes.io/projected/ef105bad-6763-4234-b52b-6d2820d48b02-kube-api-access-skdq4\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075924 4907 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075936 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075948 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ef105bad-6763-4234-b52b-6d2820d48b02-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.075958 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef105bad-6763-4234-b52b-6d2820d48b02-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.572404 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-hv2mf" event={"ID":"ef105bad-6763-4234-b52b-6d2820d48b02","Type":"ContainerDied","Data":"d1248fb6962eeda0ac2a1465b0501825fb69861c03d06ab1f3e3c6b244ab9af9"} Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.572444 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1248fb6962eeda0ac2a1465b0501825fb69861c03d06ab1f3e3c6b244ab9af9" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.572453 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-hv2mf" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.923230 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:41:24 crc kubenswrapper[4907]: E0313 15:41:24.924101 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" containerName="cinder-db-sync" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.924125 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" containerName="cinder-db-sync" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.924413 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" containerName="cinder-db-sync" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.925668 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:24 crc kubenswrapper[4907]: I0313 15:41:24.933743 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.106254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.106777 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.106979 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.107100 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.107208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.183582 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.185242 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.189712 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.190155 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.190320 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.191033 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-v5jht" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.209126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.210675 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.216409 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.216543 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.216739 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.217295 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.217511 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.217954 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.218568 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.221408 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.249373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"dnsmasq-dns-6d59c8ff75-hm7rf\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.261499 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.320390 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.320802 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.320978 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321362 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321536 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.321699 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423207 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423267 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423292 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423341 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423378 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423401 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423454 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.423799 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.424456 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.428628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.428946 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.440868 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.442584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.442612 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.503262 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:41:25 crc kubenswrapper[4907]: I0313 15:41:25.841546 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.023449 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:26 crc kubenswrapper[4907]: W0313 15:41:26.024326 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12616424_013a_47c8_9ed3_b407791537fe.slice/crio-32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896 WatchSource:0}: Error finding container 32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896: Status 404 returned error can't find the container with id 32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896 Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.599040 4907 generic.go:334] "Generic (PLEG): container finished" podID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerID="3d3465a0228af3b611e5ca1d2d283d5739801f17bcfa78429f88dbe78da94add" exitCode=0 Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.599350 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerDied","Data":"3d3465a0228af3b611e5ca1d2d283d5739801f17bcfa78429f88dbe78da94add"} Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.599382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerStarted","Data":"062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908"} Mar 13 15:41:26 crc kubenswrapper[4907]: I0313 15:41:26.602326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerStarted","Data":"32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.610863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerStarted","Data":"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.611393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerStarted","Data":"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.611414 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.613189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerStarted","Data":"04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a"} Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.613334 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.630393 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.630372634 podStartE2EDuration="2.630372634s" podCreationTimestamp="2026-03-13 15:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:27.629083838 +0000 UTC m=+5786.528871527" watchObservedRunningTime="2026-03-13 15:41:27.630372634 +0000 UTC m=+5786.530160323" Mar 13 15:41:27 crc kubenswrapper[4907]: I0313 15:41:27.652098 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" podStartSLOduration=3.6520765170000002 podStartE2EDuration="3.652076517s" podCreationTimestamp="2026-03-13 15:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:27.642833534 +0000 UTC m=+5786.542621223" watchObservedRunningTime="2026-03-13 15:41:27.652076517 +0000 UTC m=+5786.551864206" Mar 13 15:41:30 crc kubenswrapper[4907]: I0313 15:41:30.783865 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:30 crc kubenswrapper[4907]: E0313 15:41:30.784528 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.263096 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.336477 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.336742 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" containerID="cri-o://73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd" gracePeriod=10 Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.685728 4907 generic.go:334] "Generic (PLEG): container finished" podID="a122794c-3b28-4344-9817-f05a9e663110" containerID="73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd" exitCode=0 Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.686033 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerDied","Data":"73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd"} Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.868487 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.948374 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.948429 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.948451 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.949150 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.949206 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") pod \"a122794c-3b28-4344-9817-f05a9e663110\" (UID: \"a122794c-3b28-4344-9817-f05a9e663110\") " Mar 13 15:41:35 crc kubenswrapper[4907]: I0313 15:41:35.958173 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn" (OuterVolumeSpecName: "kube-api-access-d2fkn") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "kube-api-access-d2fkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.051082 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2fkn\" (UniqueName: \"kubernetes.io/projected/a122794c-3b28-4344-9817-f05a9e663110-kube-api-access-d2fkn\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.052078 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.138261 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.153978 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.154011 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.168648 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.173314 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config" (OuterVolumeSpecName: "config") pod "a122794c-3b28-4344-9817-f05a9e663110" (UID: "a122794c-3b28-4344-9817-f05a9e663110"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.255724 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.255753 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a122794c-3b28-4344-9817-f05a9e663110-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.697032 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" event={"ID":"a122794c-3b28-4344-9817-f05a9e663110","Type":"ContainerDied","Data":"8f46474c0bd8af139725a1ba39a29f7abd9cd2e12d0840cb8535991ab1cd685e"} Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.697083 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.697098 4907 scope.go:117] "RemoveContainer" containerID="73bbc4d60c70852afe1d527317bf4861e321326bd2af88b210073cad5a3608bd" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.722944 4907 scope.go:117] "RemoveContainer" containerID="ec3db5cd0d09990bedd2c273757078f6b84cc7be565b5f045a0e95e3abe11b83" Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.732089 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:36 crc kubenswrapper[4907]: I0313 15:41:36.749195 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f9699695f-pwxvd"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.104037 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.104904 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.113698 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.114034 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" containerID="cri-o://ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.129946 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.130255 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" containerID="cri-o://697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.130423 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" containerID="cri-o://da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.142147 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.142759 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" containerID="cri-o://6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.143025 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" containerID="cri-o://ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.157067 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.157296 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" containerID="cri-o://4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.190735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.190959 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" containerID="cri-o://716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" gracePeriod=30 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.330609 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.105:6080/vnc_lite.html\": dial tcp 10.217.1.105:6080: connect: connection refused" Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.714981 4907 generic.go:334] "Generic (PLEG): container finished" podID="37198210-23ed-4025-8cd9-e284c6573318" containerID="6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f" exitCode=143 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.715056 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerDied","Data":"6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f"} Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.719500 4907 generic.go:334] "Generic (PLEG): container finished" podID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" exitCode=143 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.719586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerDied","Data":"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358"} Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.721045 4907 generic.go:334] "Generic (PLEG): container finished" podID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerID="2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610" exitCode=0 Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.721074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerDied","Data":"2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610"} Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.796379 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a122794c-3b28-4344-9817-f05a9e663110" path="/var/lib/kubelet/pods/a122794c-3b28-4344-9817-f05a9e663110/volumes" Mar 13 15:41:37 crc kubenswrapper[4907]: I0313 15:41:37.937358 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.071967 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.193441 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") pod \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.193592 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") pod \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.193752 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") pod \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\" (UID: \"6244358c-3ba3-41fa-bf94-0b2bfb6ef956\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.230550 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data" (OuterVolumeSpecName: "config-data") pod "6244358c-3ba3-41fa-bf94-0b2bfb6ef956" (UID: "6244358c-3ba3-41fa-bf94-0b2bfb6ef956"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.230694 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6244358c-3ba3-41fa-bf94-0b2bfb6ef956" (UID: "6244358c-3ba3-41fa-bf94-0b2bfb6ef956"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.231046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr" (OuterVolumeSpecName: "kube-api-access-xl2nr") pod "6244358c-3ba3-41fa-bf94-0b2bfb6ef956" (UID: "6244358c-3ba3-41fa-bf94-0b2bfb6ef956"). InnerVolumeSpecName "kube-api-access-xl2nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.305549 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.305581 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.305605 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl2nr\" (UniqueName: \"kubernetes.io/projected/6244358c-3ba3-41fa-bf94-0b2bfb6ef956-kube-api-access-xl2nr\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.450097 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.534195 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.536632 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.538078 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.538120 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.633531 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") pod \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.633694 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") pod \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.633743 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") pod \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\" (UID: \"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072\") " Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.640715 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf" (OuterVolumeSpecName: "kube-api-access-47msf") pod "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" (UID: "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072"). InnerVolumeSpecName "kube-api-access-47msf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.670640 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" (UID: "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.677067 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data" (OuterVolumeSpecName: "config-data") pod "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" (UID: "d8d0d2c8-d0f8-4dc1-b65e-e694571a6072"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.733479 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.743698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6244358c-3ba3-41fa-bf94-0b2bfb6ef956","Type":"ContainerDied","Data":"97ff1fdcfd5fd91a658ff67b990106cf5c21eac69b149055f325885b224d558e"} Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748838 4907 scope.go:117] "RemoveContainer" containerID="2ff24611962f784f04b97d2b6de0e1621289a9ce82da636a0133b47ec0d27610" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748932 4907 generic.go:334] "Generic (PLEG): container finished" podID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" exitCode=0 Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748509 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerDied","Data":"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4"} Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.748994 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.749006 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d8d0d2c8-d0f8-4dc1-b65e-e694571a6072","Type":"ContainerDied","Data":"cbddd5aedb9a61c15acca2a6a299e779028d30303050e66d3356b0d2ec6abc1d"} Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.749010 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47msf\" (UniqueName: \"kubernetes.io/projected/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072-kube-api-access-47msf\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.749087 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.812942 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.825197 4907 scope.go:117] "RemoveContainer" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.840084 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.857152 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.868469 4907 scope.go:117] "RemoveContainer" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.869913 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4\": container with ID starting with 716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4 not found: ID does not exist" containerID="716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.869977 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4"} err="failed to get container status \"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4\": rpc error: code = NotFound desc = could not find container \"716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4\": container with ID starting with 716a10e355d921d8594fa1927d76d58c79889f5d1c1505d7df88ab1927fd9db4 not found: ID does not exist" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.899940 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.917627 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918139 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918166 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918196 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="init" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918204 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="init" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918220 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918228 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" Mar 13 15:41:38 crc kubenswrapper[4907]: E0313 15:41:38.918263 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918271 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918465 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" containerName="nova-cell1-conductor-conductor" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918495 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.918509 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.919267 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.934734 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.940071 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.941352 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.945379 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.951490 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952697 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952759 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952784 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952842 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952891 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h69pj\" (UniqueName: \"kubernetes.io/projected/317b6fd4-e642-4858-bc3d-2cce128ea0e7-kube-api-access-h69pj\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.952915 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:38 crc kubenswrapper[4907]: I0313 15:41:38.969701 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054673 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054779 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054902 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h69pj\" (UniqueName: \"kubernetes.io/projected/317b6fd4-e642-4858-bc3d-2cce128ea0e7-kube-api-access-h69pj\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.054935 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.061690 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.061711 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.068316 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b6fd4-e642-4858-bc3d-2cce128ea0e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.071153 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.071452 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h69pj\" (UniqueName: \"kubernetes.io/projected/317b6fd4-e642-4858-bc3d-2cce128ea0e7-kube-api-access-h69pj\") pod \"nova-cell1-novncproxy-0\" (UID: \"317b6fd4-e642-4858-bc3d-2cce128ea0e7\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.079843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.237081 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:39 crc kubenswrapper[4907]: I0313 15:41:39.261567 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:39.769022 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:39.808311 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6244358c-3ba3-41fa-bf94-0b2bfb6ef956" path="/var/lib/kubelet/pods/6244358c-3ba3-41fa-bf94-0b2bfb6ef956/volumes" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:39.812999 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d0d2c8-d0f8-4dc1-b65e-e694571a6072" path="/var/lib/kubelet/pods/d8d0d2c8-d0f8-4dc1-b65e-e694571a6072/volumes" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.522034 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.795388 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804220 4907 generic.go:334] "Generic (PLEG): container finished" podID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" exitCode=0 Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804286 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerDied","Data":"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804312 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"81ef3cd7-e06f-49f8-bc01-41b32fbb5546","Type":"ContainerDied","Data":"c85aa3e7cd1bf81f73dff61921375db95cb273392a8693e2f1fe5ad2b1d643c9"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.804328 4907 scope.go:117] "RemoveContainer" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.806918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"317b6fd4-e642-4858-bc3d-2cce128ea0e7","Type":"ContainerStarted","Data":"70f28ec9e0de121bbf148130d45c1213b4aeefcb940ccc1c2c8bb04c732a2477"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.806952 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"317b6fd4-e642-4858-bc3d-2cce128ea0e7","Type":"ContainerStarted","Data":"a5faa6bbb08bed4fc6f210842e67428795ae39508c3318118605a1b438f13212"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.818495 4907 generic.go:334] "Generic (PLEG): container finished" podID="37198210-23ed-4025-8cd9-e284c6573318" containerID="ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb" exitCode=0 Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.818558 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerDied","Data":"ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.819495 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerStarted","Data":"4c8aa8b2ae3cb083bba62c66c737dee676b8b626c0c8f53960f20765d91bb7b2"} Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.832213 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f9699695f-pwxvd" podUID="a122794c-3b28-4344-9817-f05a9e663110" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.115:5353: i/o timeout" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.845011 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.84499128 podStartE2EDuration="2.84499128s" podCreationTimestamp="2026-03-13 15:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:40.83763362 +0000 UTC m=+5799.737421309" watchObservedRunningTime="2026-03-13 15:41:40.84499128 +0000 UTC m=+5799.744778969" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.919748 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.919924 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.919978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.920036 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") pod \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\" (UID: \"81ef3cd7-e06f-49f8-bc01-41b32fbb5546\") " Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.922294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs" (OuterVolumeSpecName: "logs") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.928060 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78" (OuterVolumeSpecName: "kube-api-access-xct78") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "kube-api-access-xct78". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.940411 4907 scope.go:117] "RemoveContainer" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.946219 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.950345 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data" (OuterVolumeSpecName: "config-data") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:40 crc kubenswrapper[4907]: I0313 15:41:40.965033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81ef3cd7-e06f-49f8-bc01-41b32fbb5546" (UID: "81ef3cd7-e06f-49f8-bc01-41b32fbb5546"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.018236 4907 scope.go:117] "RemoveContainer" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.019074 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254\": container with ID starting with da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254 not found: ID does not exist" containerID="da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.019120 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254"} err="failed to get container status \"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254\": rpc error: code = NotFound desc = could not find container \"da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254\": container with ID starting with da08a8ab4ad1ce90fd28c0cf97955538dcb49b21c61a30f6afb435225b153254 not found: ID does not exist" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.019149 4907 scope.go:117] "RemoveContainer" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.020544 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358\": container with ID starting with 697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358 not found: ID does not exist" containerID="697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.020578 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358"} err="failed to get container status \"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358\": rpc error: code = NotFound desc = could not find container \"697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358\": container with ID starting with 697da8faed8339f44e9c8f069a46c98c95d40fcca3ba8965d84a7f51ddc0a358 not found: ID does not exist" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022384 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022544 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.022703 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") pod \"37198210-23ed-4025-8cd9-e284c6573318\" (UID: \"37198210-23ed-4025-8cd9-e284c6573318\") " Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023221 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023247 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xct78\" (UniqueName: \"kubernetes.io/projected/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-kube-api-access-xct78\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023260 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.023271 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81ef3cd7-e06f-49f8-bc01-41b32fbb5546-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.025106 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs" (OuterVolumeSpecName: "logs") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.031160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j" (OuterVolumeSpecName: "kube-api-access-bdf6j") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "kube-api-access-bdf6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.042414 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.054094 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data" (OuterVolumeSpecName: "config-data") pod "37198210-23ed-4025-8cd9-e284c6573318" (UID: "37198210-23ed-4025-8cd9-e284c6573318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.072025 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.073444 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.079137 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.079200 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125154 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37198210-23ed-4025-8cd9-e284c6573318-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125196 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125211 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdf6j\" (UniqueName: \"kubernetes.io/projected/37198210-23ed-4025-8cd9-e284c6573318-kube-api-access-bdf6j\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.125223 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37198210-23ed-4025-8cd9-e284c6573318-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.833280 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.834679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"37198210-23ed-4025-8cd9-e284c6573318","Type":"ContainerDied","Data":"54a994af9272c8e084ae758a74263408bf4549d1118439e6fd32c7e4d5aaa9e5"} Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.834723 4907 scope.go:117] "RemoveContainer" containerID="ae907961f503a65c4dfb28c60282669570b0349aea1f6830f2e1f6e44d773cdb" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.834844 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.843332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerStarted","Data":"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d"} Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.843537 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.869614 4907 scope.go:117] "RemoveContainer" containerID="6dcdbcfacbeb1b7e6dcbc3fb09e947521ce095a95f3551396c6c38c4e6a8b77f" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.875775 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.875751232 podStartE2EDuration="3.875751232s" podCreationTimestamp="2026-03-13 15:41:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:41.863245761 +0000 UTC m=+5800.763033440" watchObservedRunningTime="2026-03-13 15:41:41.875751232 +0000 UTC m=+5800.775538921" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.891750 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.918953 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.929172 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.953952 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954332 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954348 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954369 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954376 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954397 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954404 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" Mar 13 15:41:41 crc kubenswrapper[4907]: E0313 15:41:41.954412 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954419 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954579 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954593 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-metadata" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954604 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="37198210-23ed-4025-8cd9-e284c6573318" containerName="nova-metadata-log" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.954614 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" containerName="nova-api-api" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.955480 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.961268 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.967990 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:41 crc kubenswrapper[4907]: I0313 15:41:41.979102 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.006300 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.008472 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.011297 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.018941 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041213 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041371 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041455 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041516 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041560 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.041590 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143659 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143721 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143776 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143913 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143960 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.143996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.144019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.144443 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.144841 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.148620 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.149658 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.169672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.173626 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.173766 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"nova-api-0\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.207520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"nova-metadata-0\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.274259 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.327898 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.824949 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.855512 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerStarted","Data":"55a13d8a20cd160563d8a2c5f472cc7693d88cf83045ad53675342ed80dab573"} Mar 13 15:41:42 crc kubenswrapper[4907]: I0313 15:41:42.965354 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.523131 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.524534 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.529640 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 15:41:43 crc kubenswrapper[4907]: E0313 15:41:43.529705 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.794610 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37198210-23ed-4025-8cd9-e284c6573318" path="/var/lib/kubelet/pods/37198210-23ed-4025-8cd9-e284c6573318/volumes" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.795305 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81ef3cd7-e06f-49f8-bc01-41b32fbb5546" path="/var/lib/kubelet/pods/81ef3cd7-e06f-49f8-bc01-41b32fbb5546/volumes" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.865039 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerStarted","Data":"19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.865109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerStarted","Data":"fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.865507 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerStarted","Data":"6c6eae74def50447208be0f6d41599241acef5069efd81dc6d9fc3a28d2dd298"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.867523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerStarted","Data":"047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.867554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerStarted","Data":"c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d"} Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.885007 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.884986958 podStartE2EDuration="2.884986958s" podCreationTimestamp="2026-03-13 15:41:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:43.883443906 +0000 UTC m=+5802.783231595" watchObservedRunningTime="2026-03-13 15:41:43.884986958 +0000 UTC m=+5802.784774647" Mar 13 15:41:43 crc kubenswrapper[4907]: I0313 15:41:43.922080 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.922056321 podStartE2EDuration="2.922056321s" podCreationTimestamp="2026-03-13 15:41:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:43.908513371 +0000 UTC m=+5802.808301070" watchObservedRunningTime="2026-03-13 15:41:43.922056321 +0000 UTC m=+5802.821844010" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.237606 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.810544 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884516 4907 generic.go:334] "Generic (PLEG): container finished" podID="a8047522-cc05-44a3-b57d-2362e6457d49" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" exitCode=0 Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884561 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerDied","Data":"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34"} Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884587 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a8047522-cc05-44a3-b57d-2362e6457d49","Type":"ContainerDied","Data":"78d4b76c3a1fd63fabbe50a18b7db60eba6e1e2f95e60b44b880d052c99e1044"} Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.884624 4907 scope.go:117] "RemoveContainer" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.913691 4907 scope.go:117] "RemoveContainer" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" Mar 13 15:41:44 crc kubenswrapper[4907]: E0313 15:41:44.914233 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34\": container with ID starting with 4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34 not found: ID does not exist" containerID="4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.914263 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34"} err="failed to get container status \"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34\": rpc error: code = NotFound desc = could not find container \"4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34\": container with ID starting with 4efeb4b84a961f5a72c12fc7db6d3a11b565d9a8fc5e00e1bd6bd862f0227c34 not found: ID does not exist" Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.997929 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") pod \"a8047522-cc05-44a3-b57d-2362e6457d49\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " Mar 13 15:41:44 crc kubenswrapper[4907]: I0313 15:41:44.999949 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") pod \"a8047522-cc05-44a3-b57d-2362e6457d49\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.000051 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") pod \"a8047522-cc05-44a3-b57d-2362e6457d49\" (UID: \"a8047522-cc05-44a3-b57d-2362e6457d49\") " Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.004310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk" (OuterVolumeSpecName: "kube-api-access-77rtk") pod "a8047522-cc05-44a3-b57d-2362e6457d49" (UID: "a8047522-cc05-44a3-b57d-2362e6457d49"). InnerVolumeSpecName "kube-api-access-77rtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.028553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data" (OuterVolumeSpecName: "config-data") pod "a8047522-cc05-44a3-b57d-2362e6457d49" (UID: "a8047522-cc05-44a3-b57d-2362e6457d49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.029091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8047522-cc05-44a3-b57d-2362e6457d49" (UID: "a8047522-cc05-44a3-b57d-2362e6457d49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.102517 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.102838 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8047522-cc05-44a3-b57d-2362e6457d49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.102849 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77rtk\" (UniqueName: \"kubernetes.io/projected/a8047522-cc05-44a3-b57d-2362e6457d49-kube-api-access-77rtk\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.221949 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.235838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.245709 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: E0313 15:41:45.246129 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.246147 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.246311 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" containerName="nova-cell0-conductor-conductor" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.246969 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.253219 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.280691 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.408261 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.409208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.409264 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.510928 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.510981 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.511030 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.514951 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.519530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.528252 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"nova-cell0-conductor-0\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.609628 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.782995 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:45 crc kubenswrapper[4907]: E0313 15:41:45.783541 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:45 crc kubenswrapper[4907]: I0313 15:41:45.812990 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8047522-cc05-44a3-b57d-2362e6457d49" path="/var/lib/kubelet/pods/a8047522-cc05-44a3-b57d-2362e6457d49/volumes" Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.169020 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.907660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerStarted","Data":"c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe"} Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.908003 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerStarted","Data":"b4d79e83e6b122d073a9e884515b0623866f06609025b076fcf8513ad8311abc"} Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.908039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:46 crc kubenswrapper[4907]: I0313 15:41:46.933192 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.9331698400000001 podStartE2EDuration="1.93316984s" podCreationTimestamp="2026-03-13 15:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:46.927247718 +0000 UTC m=+5805.827035407" watchObservedRunningTime="2026-03-13 15:41:46.93316984 +0000 UTC m=+5805.832957539" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.556479 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.751019 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") pod \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.751189 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") pod \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.751310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") pod \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\" (UID: \"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30\") " Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.764130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl" (OuterVolumeSpecName: "kube-api-access-p7hsl") pod "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" (UID: "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30"). InnerVolumeSpecName "kube-api-access-p7hsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.776373 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data" (OuterVolumeSpecName: "config-data") pod "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" (UID: "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.778825 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" (UID: "61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.853027 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.853065 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.853079 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7hsl\" (UniqueName: \"kubernetes.io/projected/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30-kube-api-access-p7hsl\") on node \"crc\" DevicePath \"\"" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.915986 4907 generic.go:334] "Generic (PLEG): container finished" podID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" exitCode=0 Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916049 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916049 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerDied","Data":"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469"} Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916113 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30","Type":"ContainerDied","Data":"605697270f0438aa6a3971d720e6b66130d7e9f9c333f5d331abb06ba749d725"} Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.916142 4907 scope.go:117] "RemoveContainer" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.946178 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.947275 4907 scope.go:117] "RemoveContainer" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" Mar 13 15:41:47 crc kubenswrapper[4907]: E0313 15:41:47.947701 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469\": container with ID starting with ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469 not found: ID does not exist" containerID="ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.947730 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469"} err="failed to get container status \"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469\": rpc error: code = NotFound desc = could not find container \"ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469\": container with ID starting with ce84852f1d320a52f408594e4b89f2cbf753717a67ce179872f364593d90c469 not found: ID does not exist" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.962070 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.970028 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:47 crc kubenswrapper[4907]: E0313 15:41:47.970612 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.970636 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.970919 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" containerName="nova-scheduler-scheduler" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.971862 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.974182 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 15:41:47 crc kubenswrapper[4907]: I0313 15:41:47.978700 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.162491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.162667 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.162969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.264460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.264809 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.264893 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.268583 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.275596 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.297417 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"nova-scheduler-0\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.302424 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.795769 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 15:41:48 crc kubenswrapper[4907]: I0313 15:41:48.925220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerStarted","Data":"bba3814d9581f3c03937b4a23ce7f341cbe05c058698e236d16b4dfa33146288"} Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.237855 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.248989 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.298268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.792461 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30" path="/var/lib/kubelet/pods/61a2ddd9-1fcc-427e-aaa2-f092ebb7bd30/volumes" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.939028 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerStarted","Data":"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88"} Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.950023 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 15:41:49 crc kubenswrapper[4907]: I0313 15:41:49.965702 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.965687213 podStartE2EDuration="2.965687213s" podCreationTimestamp="2026-03-13 15:41:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:49.962342092 +0000 UTC m=+5808.862129771" watchObservedRunningTime="2026-03-13 15:41:49.965687213 +0000 UTC m=+5808.865474902" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.275170 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.275519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.328942 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:41:52 crc kubenswrapper[4907]: I0313 15:41:52.330504 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.302941 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.358183 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.123:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.440082 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.124:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.440196 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.124:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:53 crc kubenswrapper[4907]: I0313 15:41:53.440083 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.123:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.651769 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.713046 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.715349 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.731270 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.765830 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.905206 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.905275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.905313 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.906006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.906234 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:55 crc kubenswrapper[4907]: I0313 15:41:55.906380 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008176 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008255 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008369 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.008423 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.013596 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.015385 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.017370 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.027463 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.028534 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"cinder-scheduler-0\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.063470 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.667212 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:41:56 crc kubenswrapper[4907]: W0313 15:41:56.672179 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfe30c39_e9da_49f4_a9da_ebbce147b863.slice/crio-36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed WatchSource:0}: Error finding container 36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed: Status 404 returned error can't find the container with id 36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.782935 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:41:56 crc kubenswrapper[4907]: E0313 15:41:56.783553 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.988084 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.988332 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" containerID="cri-o://f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" gracePeriod=30 Mar 13 15:41:56 crc kubenswrapper[4907]: I0313 15:41:56.988439 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" containerID="cri-o://b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" gracePeriod=30 Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.035710 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerStarted","Data":"36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed"} Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.477845 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.479759 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.481519 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.527759 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553091 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553115 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553256 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553329 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553368 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553507 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4d56\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-kube-api-access-z4d56\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.553954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554012 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554115 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-run\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554182 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554240 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.554267 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655512 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-run\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655574 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655606 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655629 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655657 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655687 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655713 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655709 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-run\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655746 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655872 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.655959 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656299 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4d56\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-kube-api-access-z4d56\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656484 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656541 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656634 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.656778 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-dev\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-sys\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657552 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.657683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.658126 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/697762c8-b121-410a-a5db-21916ca1791f-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660678 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.660691 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.664378 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/697762c8-b121-410a-a5db-21916ca1791f-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.674070 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4d56\" (UniqueName: \"kubernetes.io/projected/697762c8-b121-410a-a5db-21916ca1791f-kube-api-access-z4d56\") pod \"cinder-volume-volume1-0\" (UID: \"697762c8-b121-410a-a5db-21916ca1791f\") " pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:57 crc kubenswrapper[4907]: I0313 15:41:57.833987 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.074091 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerStarted","Data":"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6"} Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.074698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerStarted","Data":"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27"} Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.080579 4907 generic.go:334] "Generic (PLEG): container finished" podID="12616424-013a-47c8-9ed3-b407791537fe" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" exitCode=143 Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.080633 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerDied","Data":"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b"} Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.107183 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.107165302 podStartE2EDuration="3.107165302s" podCreationTimestamp="2026-03-13 15:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:41:58.100084059 +0000 UTC m=+5816.999871748" watchObservedRunningTime="2026-03-13 15:41:58.107165302 +0000 UTC m=+5817.006952991" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.302730 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.327901 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.329700 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.335999 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.353554 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.365835 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.373700 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-sys\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.373774 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.373843 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-run\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-dev\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374162 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-ceph\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kls8z\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-kube-api-access-kls8z\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374332 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374403 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-scripts\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374485 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374746 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.374819 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375132 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375273 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375349 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-lib-modules\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.375412 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478517 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kls8z\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-kube-api-access-kls8z\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478560 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478580 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-scripts\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478601 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478617 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478654 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478676 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478716 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478733 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-lib-modules\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478749 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478810 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-sys\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478825 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478846 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-run\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478866 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-dev\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.478894 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-ceph\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.479341 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.482794 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484777 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484848 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-sys\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484912 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-lib-modules\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.484979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.485177 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-run\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.485244 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.486667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-ceph\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.486735 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-dev\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.486788 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d92d5136-2169-44cc-9900-c3335a4b69eb-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.489680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.499659 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-scripts\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.503443 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.503566 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kls8z\" (UniqueName: \"kubernetes.io/projected/d92d5136-2169-44cc-9900-c3335a4b69eb-kube-api-access-kls8z\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.506390 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d92d5136-2169-44cc-9900-c3335a4b69eb-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"d92d5136-2169-44cc-9900-c3335a4b69eb\") " pod="openstack/cinder-backup-0" Mar 13 15:41:58 crc kubenswrapper[4907]: I0313 15:41:58.655958 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Mar 13 15:41:59 crc kubenswrapper[4907]: I0313 15:41:59.090794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"697762c8-b121-410a-a5db-21916ca1791f","Type":"ContainerStarted","Data":"dd76292c1d3583e1de82c7463dec626e3e766b926d7d02c412e867389a2f86d2"} Mar 13 15:41:59 crc kubenswrapper[4907]: I0313 15:41:59.117662 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 15:41:59 crc kubenswrapper[4907]: I0313 15:41:59.255762 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.099178 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"697762c8-b121-410a-a5db-21916ca1791f","Type":"ContainerStarted","Data":"67e59dc847d580867e158938a2fb3b51424233c6ff48024374cf07e953a69756"} Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.099809 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"697762c8-b121-410a-a5db-21916ca1791f","Type":"ContainerStarted","Data":"bd5431dfa5b4119b17ac9f6b0b76ab0c54b5f6b19e74d192b265bb7cb3e2db8a"} Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.103659 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d92d5136-2169-44cc-9900-c3335a4b69eb","Type":"ContainerStarted","Data":"14a65b0f6bf8a07e477af765e4bdab5239237933ab32aa4a7cef322ba35a3dc1"} Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.126347 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.13408908 podStartE2EDuration="3.126328539s" podCreationTimestamp="2026-03-13 15:41:57 +0000 UTC" firstStartedPulling="2026-03-13 15:41:58.485773647 +0000 UTC m=+5817.385561336" lastFinishedPulling="2026-03-13 15:41:59.478013106 +0000 UTC m=+5818.377800795" observedRunningTime="2026-03-13 15:42:00.124643643 +0000 UTC m=+5819.024431352" watchObservedRunningTime="2026-03-13 15:42:00.126328539 +0000 UTC m=+5819.026116228" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.150139 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.151725 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.156993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.157477 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.157477 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.157844 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.274421 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.274482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.328463 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.328811 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.329852 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"auto-csr-approver-29556942-mdddr\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.432301 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"auto-csr-approver-29556942-mdddr\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.455461 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"auto-csr-approver-29556942-mdddr\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.482996 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.660768 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844747 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844825 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.844975 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.845087 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.845119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") pod \"12616424-013a-47c8-9ed3-b407791537fe\" (UID: \"12616424-013a-47c8-9ed3-b407791537fe\") " Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.845802 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs" (OuterVolumeSpecName: "logs") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.848661 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.851970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk" (OuterVolumeSpecName: "kube-api-access-dh7tk") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "kube-api-access-dh7tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.852092 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.857224 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts" (OuterVolumeSpecName: "scripts") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.904423 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.923499 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data" (OuterVolumeSpecName: "config-data") pod "12616424-013a-47c8-9ed3-b407791537fe" (UID: "12616424-013a-47c8-9ed3-b407791537fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947678 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12616424-013a-47c8-9ed3-b407791537fe-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947712 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947723 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947731 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947740 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12616424-013a-47c8-9ed3-b407791537fe-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947747 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12616424-013a-47c8-9ed3-b407791537fe-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:00 crc kubenswrapper[4907]: I0313 15:42:00.947758 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh7tk\" (UniqueName: \"kubernetes.io/projected/12616424-013a-47c8-9ed3-b407791537fe-kube-api-access-dh7tk\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.059761 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.065038 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116619 4907 generic.go:334] "Generic (PLEG): container finished" podID="12616424-013a-47c8-9ed3-b407791537fe" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" exitCode=0 Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116680 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerDied","Data":"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116707 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"12616424-013a-47c8-9ed3-b407791537fe","Type":"ContainerDied","Data":"32e9d54b2cd43cd1f900bf7ababaaace7fcabaccb1a6e6993088fcc8c885c896"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116723 4907 scope.go:117] "RemoveContainer" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.116837 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.135718 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556942-mdddr" event={"ID":"ed0ed6c7-e5d4-4770-a9aa-308c0853573e","Type":"ContainerStarted","Data":"763973579a2d2ec12ba365a5be5b38596e6d77a76e685b5bbe27b06cc46b881e"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.140423 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d92d5136-2169-44cc-9900-c3335a4b69eb","Type":"ContainerStarted","Data":"29454c8b38fc418834a59d22c32579d841bc25e689c84de54872ee146876dd55"} Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.189045 4907 scope.go:117] "RemoveContainer" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.197167 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.207128 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.225782 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.226733 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.226772 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.226792 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.226800 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.227064 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api-log" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.227094 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.228399 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.234942 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.242510 4907 scope.go:117] "RemoveContainer" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.247450 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717\": container with ID starting with b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717 not found: ID does not exist" containerID="b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.247487 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717"} err="failed to get container status \"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717\": rpc error: code = NotFound desc = could not find container \"b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717\": container with ID starting with b5b4ab9a472db32a9c0ae91b2509ae28f1254ad915d9737b1279509e0c8d4717 not found: ID does not exist" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.247509 4907 scope.go:117] "RemoveContainer" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" Mar 13 15:42:01 crc kubenswrapper[4907]: E0313 15:42:01.249901 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b\": container with ID starting with f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b not found: ID does not exist" containerID="f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.249929 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b"} err="failed to get container status \"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b\": rpc error: code = NotFound desc = could not find container \"f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b\": container with ID starting with f234443780f86c06ddc613fb1859ac940ed1d9617b56539195ffdc071bbc882b not found: ID does not exist" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.254204 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.264711 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265394 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c82a596-df2b-4e69-abe9-5f4662a3c6da-logs\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265656 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-scripts\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265730 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.265760 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c82a596-df2b-4e69-abe9-5f4662a3c6da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.266238 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data-custom\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.266303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf67l\" (UniqueName: \"kubernetes.io/projected/8c82a596-df2b-4e69-abe9-5f4662a3c6da-kube-api-access-mf67l\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.371215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data-custom\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.371287 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf67l\" (UniqueName: \"kubernetes.io/projected/8c82a596-df2b-4e69-abe9-5f4662a3c6da-kube-api-access-mf67l\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.372521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.372698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c82a596-df2b-4e69-abe9-5f4662a3c6da-logs\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.374458 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-scripts\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.374502 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.374645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c82a596-df2b-4e69-abe9-5f4662a3c6da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.376991 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8c82a596-df2b-4e69-abe9-5f4662a3c6da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.378193 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c82a596-df2b-4e69-abe9-5f4662a3c6da-logs\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.382861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-scripts\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.396625 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf67l\" (UniqueName: \"kubernetes.io/projected/8c82a596-df2b-4e69-abe9-5f4662a3c6da-kube-api-access-mf67l\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.396950 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data-custom\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.398019 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.402112 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c82a596-df2b-4e69-abe9-5f4662a3c6da-config-data\") pod \"cinder-api-0\" (UID: \"8c82a596-df2b-4e69-abe9-5f4662a3c6da\") " pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.616415 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 15:42:01 crc kubenswrapper[4907]: I0313 15:42:01.811482 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12616424-013a-47c8-9ed3-b407791537fe" path="/var/lib/kubelet/pods/12616424-013a-47c8-9ed3-b407791537fe/volumes" Mar 13 15:42:02 crc kubenswrapper[4907]: W0313 15:42:02.083691 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c82a596_df2b_4e69_abe9_5f4662a3c6da.slice/crio-ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7 WatchSource:0}: Error finding container ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7: Status 404 returned error can't find the container with id ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7 Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.090637 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.152816 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8c82a596-df2b-4e69-abe9-5f4662a3c6da","Type":"ContainerStarted","Data":"ea21a86a7210843a61ba8d606f21ee3219af37b3433a82cdba8aae6e4a77d1b7"} Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.159660 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"d92d5136-2169-44cc-9900-c3335a4b69eb","Type":"ContainerStarted","Data":"427089c1bbd58a37d7507fb363c5b14d82a397d44285e0e4981cb8c2241be726"} Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.204383 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.701627639 podStartE2EDuration="4.204341225s" podCreationTimestamp="2026-03-13 15:41:58 +0000 UTC" firstStartedPulling="2026-03-13 15:41:59.262118248 +0000 UTC m=+5818.161905937" lastFinishedPulling="2026-03-13 15:42:00.764831834 +0000 UTC m=+5819.664619523" observedRunningTime="2026-03-13 15:42:02.196312396 +0000 UTC m=+5821.096100085" watchObservedRunningTime="2026-03-13 15:42:02.204341225 +0000 UTC m=+5821.104128914" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.281979 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.285821 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.285970 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.332055 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.334365 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.334842 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:42:02 crc kubenswrapper[4907]: I0313 15:42:02.835411 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.171242 4907 generic.go:334] "Generic (PLEG): container finished" podID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerID="7c596675c8272c2dfe2507df7b83ab578498cef4ae7b5f27995638598d3ea046" exitCode=0 Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.171334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556942-mdddr" event={"ID":"ed0ed6c7-e5d4-4770-a9aa-308c0853573e","Type":"ContainerDied","Data":"7c596675c8272c2dfe2507df7b83ab578498cef4ae7b5f27995638598d3ea046"} Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.187693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8c82a596-df2b-4e69-abe9-5f4662a3c6da","Type":"ContainerStarted","Data":"268cc9bc4e1fb3e66dd707585b30d6d2cb836b5a40afb999a88784cf686be14c"} Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.191114 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.192636 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 15:42:03 crc kubenswrapper[4907]: I0313 15:42:03.657143 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.196219 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8c82a596-df2b-4e69-abe9-5f4662a3c6da","Type":"ContainerStarted","Data":"a7a8a2bc0d7150294189b41f6abfff6c5259b80a09c5f094de918c944f25e0b2"} Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.232392 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.232373354 podStartE2EDuration="3.232373354s" podCreationTimestamp="2026-03-13 15:42:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:42:04.21721359 +0000 UTC m=+5823.117001279" watchObservedRunningTime="2026-03-13 15:42:04.232373354 +0000 UTC m=+5823.132161043" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.555723 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.645726 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") pod \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\" (UID: \"ed0ed6c7-e5d4-4770-a9aa-308c0853573e\") " Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.651451 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh" (OuterVolumeSpecName: "kube-api-access-s8lgh") pod "ed0ed6c7-e5d4-4770-a9aa-308c0853573e" (UID: "ed0ed6c7-e5d4-4770-a9aa-308c0853573e"). InnerVolumeSpecName "kube-api-access-s8lgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:42:04 crc kubenswrapper[4907]: I0313 15:42:04.748344 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8lgh\" (UniqueName: \"kubernetes.io/projected/ed0ed6c7-e5d4-4770-a9aa-308c0853573e-kube-api-access-s8lgh\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208319 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556942-mdddr" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208340 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556942-mdddr" event={"ID":"ed0ed6c7-e5d4-4770-a9aa-308c0853573e","Type":"ContainerDied","Data":"763973579a2d2ec12ba365a5be5b38596e6d77a76e685b5bbe27b06cc46b881e"} Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208370 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="763973579a2d2ec12ba365a5be5b38596e6d77a76e685b5bbe27b06cc46b881e" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.208540 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.504967 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="12616424-013a-47c8-9ed3-b407791537fe" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.120:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.648178 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.656246 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556936-bwsf9"] Mar 13 15:42:05 crc kubenswrapper[4907]: I0313 15:42:05.796088 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7" path="/var/lib/kubelet/pods/38ab6eb6-9c07-41f4-87e4-d74cb4e5aff7/volumes" Mar 13 15:42:06 crc kubenswrapper[4907]: I0313 15:42:06.262083 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 15:42:06 crc kubenswrapper[4907]: I0313 15:42:06.302479 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:07 crc kubenswrapper[4907]: I0313 15:42:07.228705 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" containerID="cri-o://24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" gracePeriod=30 Mar 13 15:42:07 crc kubenswrapper[4907]: I0313 15:42:07.228873 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" containerID="cri-o://fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" gracePeriod=30 Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.053095 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.239606 4907 generic.go:334] "Generic (PLEG): container finished" podID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" exitCode=0 Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.239704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerDied","Data":"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6"} Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.872665 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Mar 13 15:42:08 crc kubenswrapper[4907]: I0313 15:42:08.888413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023448 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023858 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023961 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.023989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.024018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.024077 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") pod \"bfe30c39-e9da-49f4-a9da-ebbce147b863\" (UID: \"bfe30c39-e9da-49f4-a9da-ebbce147b863\") " Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.024920 4907 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe30c39-e9da-49f4-a9da-ebbce147b863-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.029152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts" (OuterVolumeSpecName: "scripts") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.033157 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84" (OuterVolumeSpecName: "kube-api-access-d5t84") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "kube-api-access-d5t84". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.042135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.087675 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126135 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126167 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126179 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.126190 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5t84\" (UniqueName: \"kubernetes.io/projected/bfe30c39-e9da-49f4-a9da-ebbce147b863-kube-api-access-d5t84\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.141184 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data" (OuterVolumeSpecName: "config-data") pod "bfe30c39-e9da-49f4-a9da-ebbce147b863" (UID: "bfe30c39-e9da-49f4-a9da-ebbce147b863"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.228796 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe30c39-e9da-49f4-a9da-ebbce147b863-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.250939 4907 generic.go:334] "Generic (PLEG): container finished" podID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" exitCode=0 Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.250992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerDied","Data":"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27"} Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.251029 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe30c39-e9da-49f4-a9da-ebbce147b863","Type":"ContainerDied","Data":"36d4f2be9f852fdbc463601ca568186a38921963adf0237624ff766e5451ebed"} Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.251049 4907 scope.go:117] "RemoveContainer" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.251329 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.291721 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.296158 4907 scope.go:117] "RemoveContainer" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.297642 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318141 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.318601 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318621 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.318650 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318657 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.318671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerName="oc" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318678 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerName="oc" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318903 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="probe" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318931 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" containerName="oc" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.318942 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" containerName="cinder-scheduler" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.320203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.325125 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.329779 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330545 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a999cb1-5df0-43ae-946f-eae3f2c53848-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330604 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwxzz\" (UniqueName: \"kubernetes.io/projected/3a999cb1-5df0-43ae-946f-eae3f2c53848-kube-api-access-wwxzz\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330686 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330704 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.330764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.358026 4907 scope.go:117] "RemoveContainer" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.359755 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6\": container with ID starting with fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6 not found: ID does not exist" containerID="fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.359808 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6"} err="failed to get container status \"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6\": rpc error: code = NotFound desc = could not find container \"fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6\": container with ID starting with fbe67e4b2a6e6d29dd61f35ad336b329273792eab74bfcf4dd2ade78056599e6 not found: ID does not exist" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.359838 4907 scope.go:117] "RemoveContainer" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" Mar 13 15:42:09 crc kubenswrapper[4907]: E0313 15:42:09.361292 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27\": container with ID starting with 24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27 not found: ID does not exist" containerID="24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.361332 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27"} err="failed to get container status \"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27\": rpc error: code = NotFound desc = could not find container \"24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27\": container with ID starting with 24af8c70da41dc6cb88e8d25d6b679b4b7c4b0986765337bf4ab3605a412ae27 not found: ID does not exist" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.432914 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433042 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwxzz\" (UniqueName: \"kubernetes.io/projected/3a999cb1-5df0-43ae-946f-eae3f2c53848-kube-api-access-wwxzz\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433085 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433108 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433195 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433277 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a999cb1-5df0-43ae-946f-eae3f2c53848-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.433494 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a999cb1-5df0-43ae-946f-eae3f2c53848-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.437260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.437526 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.438001 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.442349 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a999cb1-5df0-43ae-946f-eae3f2c53848-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.450541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwxzz\" (UniqueName: \"kubernetes.io/projected/3a999cb1-5df0-43ae-946f-eae3f2c53848-kube-api-access-wwxzz\") pod \"cinder-scheduler-0\" (UID: \"3a999cb1-5df0-43ae-946f-eae3f2c53848\") " pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.637069 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 15:42:09 crc kubenswrapper[4907]: I0313 15:42:09.796589 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe30c39-e9da-49f4-a9da-ebbce147b863" path="/var/lib/kubelet/pods/bfe30c39-e9da-49f4-a9da-ebbce147b863/volumes" Mar 13 15:42:10 crc kubenswrapper[4907]: I0313 15:42:10.054128 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 15:42:10 crc kubenswrapper[4907]: W0313 15:42:10.062349 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a999cb1_5df0_43ae_946f_eae3f2c53848.slice/crio-61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e WatchSource:0}: Error finding container 61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e: Status 404 returned error can't find the container with id 61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e Mar 13 15:42:10 crc kubenswrapper[4907]: I0313 15:42:10.260653 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a999cb1-5df0-43ae-946f-eae3f2c53848","Type":"ContainerStarted","Data":"61ecef7c54a689fe4149ac7c8e9ba61eb21ae1c4bc42a3f8e894b735b5cb869e"} Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.274597 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a999cb1-5df0-43ae-946f-eae3f2c53848","Type":"ContainerStarted","Data":"3504104e47cbc2a010fa4b4eaa305d2b337e3a5bec3d3fca454e6618c2ac801b"} Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.275099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a999cb1-5df0-43ae-946f-eae3f2c53848","Type":"ContainerStarted","Data":"bf04a7270e06252486e2cbf6625bd906eddc33b84f8b5fb6a27a97ccc0c75eaf"} Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.298170 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.298154383 podStartE2EDuration="2.298154383s" podCreationTimestamp="2026-03-13 15:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:42:11.297558508 +0000 UTC m=+5830.197346217" watchObservedRunningTime="2026-03-13 15:42:11.298154383 +0000 UTC m=+5830.197942072" Mar 13 15:42:11 crc kubenswrapper[4907]: I0313 15:42:11.786503 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:11 crc kubenswrapper[4907]: E0313 15:42:11.787113 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:13 crc kubenswrapper[4907]: I0313 15:42:13.468849 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 15:42:14 crc kubenswrapper[4907]: I0313 15:42:14.637696 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 15:42:19 crc kubenswrapper[4907]: I0313 15:42:19.822267 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 15:42:22 crc kubenswrapper[4907]: I0313 15:42:22.782550 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:22 crc kubenswrapper[4907]: E0313 15:42:22.783091 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:29 crc kubenswrapper[4907]: I0313 15:42:29.064280 4907 scope.go:117] "RemoveContainer" containerID="a02ae219ab9b5a1fd678d76ac3b1c7e41a88bec1b6ae3e7a16334b35c97e0864" Mar 13 15:42:34 crc kubenswrapper[4907]: I0313 15:42:34.782254 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:34 crc kubenswrapper[4907]: E0313 15:42:34.782844 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:49 crc kubenswrapper[4907]: I0313 15:42:49.782998 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:42:49 crc kubenswrapper[4907]: E0313 15:42:49.783842 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.485222 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.490028 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.501742 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.638512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.639030 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.639077 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.764764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.764986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.765025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.765388 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.765486 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.792986 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"redhat-marketplace-4cjst\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:54 crc kubenswrapper[4907]: I0313 15:42:54.826297 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.361207 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:42:55 crc kubenswrapper[4907]: W0313 15:42:55.364726 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24f03607_91fb_432b_afba_5d7c7634e98e.slice/crio-7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c WatchSource:0}: Error finding container 7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c: Status 404 returned error can't find the container with id 7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.691289 4907 generic.go:334] "Generic (PLEG): container finished" podID="24f03607-91fb-432b-afba-5d7c7634e98e" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" exitCode=0 Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.691397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8"} Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.691668 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerStarted","Data":"7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c"} Mar 13 15:42:55 crc kubenswrapper[4907]: I0313 15:42:55.693238 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:42:57 crc kubenswrapper[4907]: I0313 15:42:57.714363 4907 generic.go:334] "Generic (PLEG): container finished" podID="24f03607-91fb-432b-afba-5d7c7634e98e" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" exitCode=0 Mar 13 15:42:57 crc kubenswrapper[4907]: I0313 15:42:57.714540 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01"} Mar 13 15:42:58 crc kubenswrapper[4907]: I0313 15:42:58.724064 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerStarted","Data":"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588"} Mar 13 15:42:58 crc kubenswrapper[4907]: I0313 15:42:58.745686 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4cjst" podStartSLOduration=2.312450315 podStartE2EDuration="4.745661094s" podCreationTimestamp="2026-03-13 15:42:54 +0000 UTC" firstStartedPulling="2026-03-13 15:42:55.693013771 +0000 UTC m=+5874.592801450" lastFinishedPulling="2026-03-13 15:42:58.12622454 +0000 UTC m=+5877.026012229" observedRunningTime="2026-03-13 15:42:58.738740415 +0000 UTC m=+5877.638528114" watchObservedRunningTime="2026-03-13 15:42:58.745661094 +0000 UTC m=+5877.645448793" Mar 13 15:43:02 crc kubenswrapper[4907]: I0313 15:43:02.782552 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:02 crc kubenswrapper[4907]: E0313 15:43:02.783599 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:04 crc kubenswrapper[4907]: I0313 15:43:04.827467 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:04 crc kubenswrapper[4907]: I0313 15:43:04.827787 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:04 crc kubenswrapper[4907]: I0313 15:43:04.870845 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:05 crc kubenswrapper[4907]: I0313 15:43:05.835532 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:05 crc kubenswrapper[4907]: I0313 15:43:05.893349 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:43:07 crc kubenswrapper[4907]: I0313 15:43:07.811690 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4cjst" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" containerID="cri-o://ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" gracePeriod=2 Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.296093 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.471236 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") pod \"24f03607-91fb-432b-afba-5d7c7634e98e\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.471289 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") pod \"24f03607-91fb-432b-afba-5d7c7634e98e\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.471422 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") pod \"24f03607-91fb-432b-afba-5d7c7634e98e\" (UID: \"24f03607-91fb-432b-afba-5d7c7634e98e\") " Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.472469 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities" (OuterVolumeSpecName: "utilities") pod "24f03607-91fb-432b-afba-5d7c7634e98e" (UID: "24f03607-91fb-432b-afba-5d7c7634e98e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.477607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt" (OuterVolumeSpecName: "kube-api-access-lr6wt") pod "24f03607-91fb-432b-afba-5d7c7634e98e" (UID: "24f03607-91fb-432b-afba-5d7c7634e98e"). InnerVolumeSpecName "kube-api-access-lr6wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.495894 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24f03607-91fb-432b-afba-5d7c7634e98e" (UID: "24f03607-91fb-432b-afba-5d7c7634e98e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.573510 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.573553 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f03607-91fb-432b-afba-5d7c7634e98e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.573566 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr6wt\" (UniqueName: \"kubernetes.io/projected/24f03607-91fb-432b-afba-5d7c7634e98e-kube-api-access-lr6wt\") on node \"crc\" DevicePath \"\"" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827134 4907 generic.go:334] "Generic (PLEG): container finished" podID="24f03607-91fb-432b-afba-5d7c7634e98e" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" exitCode=0 Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827190 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cjst" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827200 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588"} Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827242 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cjst" event={"ID":"24f03607-91fb-432b-afba-5d7c7634e98e","Type":"ContainerDied","Data":"7965e8c5457fc504594c88a465076bcea4651faf50c777df09b4aa7e11a81d7c"} Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.827270 4907 scope.go:117] "RemoveContainer" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.849013 4907 scope.go:117] "RemoveContainer" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.869805 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.882961 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cjst"] Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.897439 4907 scope.go:117] "RemoveContainer" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.921895 4907 scope.go:117] "RemoveContainer" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" Mar 13 15:43:08 crc kubenswrapper[4907]: E0313 15:43:08.922394 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588\": container with ID starting with ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588 not found: ID does not exist" containerID="ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922426 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588"} err="failed to get container status \"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588\": rpc error: code = NotFound desc = could not find container \"ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588\": container with ID starting with ec5a19c81a6873a70bb62001169af40470c3baefad28d2afe99af6e6991ef588 not found: ID does not exist" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922448 4907 scope.go:117] "RemoveContainer" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" Mar 13 15:43:08 crc kubenswrapper[4907]: E0313 15:43:08.922834 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01\": container with ID starting with f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01 not found: ID does not exist" containerID="f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922892 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01"} err="failed to get container status \"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01\": rpc error: code = NotFound desc = could not find container \"f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01\": container with ID starting with f1e605d4688b3741b8cfb0f2118afdc7b3de9f3ec5a9a3b8e5ba6a60f2026d01 not found: ID does not exist" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.922924 4907 scope.go:117] "RemoveContainer" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" Mar 13 15:43:08 crc kubenswrapper[4907]: E0313 15:43:08.923223 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8\": container with ID starting with 5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8 not found: ID does not exist" containerID="5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8" Mar 13 15:43:08 crc kubenswrapper[4907]: I0313 15:43:08.923248 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8"} err="failed to get container status \"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8\": rpc error: code = NotFound desc = could not find container \"5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8\": container with ID starting with 5ec383916b83bb60b7b67c55360eb9d838cf4da064b8916d7f76d953c53403c8 not found: ID does not exist" Mar 13 15:43:09 crc kubenswrapper[4907]: I0313 15:43:09.795445 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" path="/var/lib/kubelet/pods/24f03607-91fb-432b-afba-5d7c7634e98e/volumes" Mar 13 15:43:15 crc kubenswrapper[4907]: I0313 15:43:15.782903 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:15 crc kubenswrapper[4907]: E0313 15:43:15.783679 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:27 crc kubenswrapper[4907]: I0313 15:43:27.782450 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:27 crc kubenswrapper[4907]: E0313 15:43:27.783304 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:38 crc kubenswrapper[4907]: I0313 15:43:38.783079 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:38 crc kubenswrapper[4907]: E0313 15:43:38.783988 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:49 crc kubenswrapper[4907]: I0313 15:43:49.787991 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:43:49 crc kubenswrapper[4907]: E0313 15:43:49.788696 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.753203 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2"] Mar 13 15:43:56 crc kubenswrapper[4907]: E0313 15:43:56.754108 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-content" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754122 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-content" Mar 13 15:43:56 crc kubenswrapper[4907]: E0313 15:43:56.754140 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-utilities" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754147 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="extract-utilities" Mar 13 15:43:56 crc kubenswrapper[4907]: E0313 15:43:56.754168 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754175 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754337 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f03607-91fb-432b-afba-5d7c7634e98e" containerName="registry-server" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.754924 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.761890 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-w5dfv" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.762336 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.772805 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-v5b57"] Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.775529 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.789421 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2"] Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.805177 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v5b57"] Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884491 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884537 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-etc-ovs\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884566 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-log\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884659 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-lib\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884698 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzxds\" (UniqueName: \"kubernetes.io/projected/1cf43c88-f066-4979-8590-b8e6668aab09-kube-api-access-gzxds\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884724 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cf43c88-f066-4979-8590-b8e6668aab09-scripts\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884750 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-scripts\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884784 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-log-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884914 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-run\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.884952 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92p4x\" (UniqueName: \"kubernetes.io/projected/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-kube-api-access-92p4x\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986379 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986431 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-etc-ovs\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-log\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986518 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-lib\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986558 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzxds\" (UniqueName: \"kubernetes.io/projected/1cf43c88-f066-4979-8590-b8e6668aab09-kube-api-access-gzxds\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cf43c88-f066-4979-8590-b8e6668aab09-scripts\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-scripts\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-log-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986680 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986707 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-run\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.986738 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92p4x\" (UniqueName: \"kubernetes.io/projected/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-kube-api-access-92p4x\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987381 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-etc-ovs\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-log\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-log-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987508 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1cf43c88-f066-4979-8590-b8e6668aab09-var-run-ovn\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987574 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-run\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.987680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-var-lib\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.989500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-scripts\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:56 crc kubenswrapper[4907]: I0313 15:43:56.989797 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1cf43c88-f066-4979-8590-b8e6668aab09-scripts\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.006850 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzxds\" (UniqueName: \"kubernetes.io/projected/1cf43c88-f066-4979-8590-b8e6668aab09-kube-api-access-gzxds\") pod \"ovn-controller-zb9k2\" (UID: \"1cf43c88-f066-4979-8590-b8e6668aab09\") " pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.015458 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92p4x\" (UniqueName: \"kubernetes.io/projected/34b897ab-ce41-4cd5-bcae-cea3c0dea9e9-kube-api-access-92p4x\") pod \"ovn-controller-ovs-v5b57\" (UID: \"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9\") " pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.075107 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.091061 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:43:57 crc kubenswrapper[4907]: I0313 15:43:57.554767 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2"] Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.062153 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-v5b57"] Mar 13 15:43:58 crc kubenswrapper[4907]: W0313 15:43:58.066324 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34b897ab_ce41_4cd5_bcae_cea3c0dea9e9.slice/crio-16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c WatchSource:0}: Error finding container 16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c: Status 404 returned error can't find the container with id 16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.335630 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerStarted","Data":"16325fb744d63b685dea9ff2553f2408a5735d291ee63bfc31b4a34d96d3664c"} Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.337193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2" event={"ID":"1cf43c88-f066-4979-8590-b8e6668aab09","Type":"ContainerStarted","Data":"666a17c67dac296199ae01f33215fddd1a37e44ab70056e8156c835ff9a3812b"} Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.337236 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2" event={"ID":"1cf43c88-f066-4979-8590-b8e6668aab09","Type":"ContainerStarted","Data":"3b7c23409163ae5d10332502c897514fd1d764456aac8e68f45261995b6934b5"} Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.337533 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zb9k2" Mar 13 15:43:58 crc kubenswrapper[4907]: I0313 15:43:58.360319 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zb9k2" podStartSLOduration=2.360303212 podStartE2EDuration="2.360303212s" podCreationTimestamp="2026-03-13 15:43:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:43:58.355567592 +0000 UTC m=+5937.255355321" watchObservedRunningTime="2026-03-13 15:43:58.360303212 +0000 UTC m=+5937.260090901" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.369967 4907 generic.go:334] "Generic (PLEG): container finished" podID="34b897ab-ce41-4cd5-bcae-cea3c0dea9e9" containerID="fecc41f131bc004354ba494350d5e8f1e25a264511e3c8936fa8c28deae60120" exitCode=0 Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.370977 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerDied","Data":"fecc41f131bc004354ba494350d5e8f1e25a264511e3c8936fa8c28deae60120"} Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.374038 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-xstlm"] Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.397331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.409318 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.422988 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xstlm"] Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444371 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvq2t\" (UniqueName: \"kubernetes.io/projected/f9398439-2570-42ec-b6fc-f9770b988c73-kube-api-access-pvq2t\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9398439-2570-42ec-b6fc-f9770b988c73-config\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444661 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovs-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.444711 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovn-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546006 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvq2t\" (UniqueName: \"kubernetes.io/projected/f9398439-2570-42ec-b6fc-f9770b988c73-kube-api-access-pvq2t\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546094 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9398439-2570-42ec-b6fc-f9770b988c73-config\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546129 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovs-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovn-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovn-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.546810 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f9398439-2570-42ec-b6fc-f9770b988c73-ovs-rundir\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.547517 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9398439-2570-42ec-b6fc-f9770b988c73-config\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.575226 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvq2t\" (UniqueName: \"kubernetes.io/projected/f9398439-2570-42ec-b6fc-f9770b988c73-kube-api-access-pvq2t\") pod \"ovn-controller-metrics-xstlm\" (UID: \"f9398439-2570-42ec-b6fc-f9770b988c73\") " pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:43:59 crc kubenswrapper[4907]: I0313 15:43:59.744605 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-xstlm" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.129651 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.130969 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.133447 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.133688 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.133912 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.147652 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.207986 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-xstlm"] Mar 13 15:44:00 crc kubenswrapper[4907]: W0313 15:44:00.214410 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9398439_2570_42ec_b6fc_f9770b988c73.slice/crio-8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3 WatchSource:0}: Error finding container 8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3: Status 404 returned error can't find the container with id 8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3 Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.257276 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"auto-csr-approver-29556944-gcp6x\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.359027 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"auto-csr-approver-29556944-gcp6x\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.378591 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"auto-csr-approver-29556944-gcp6x\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerStarted","Data":"c66b2e74b53f59150cce15506d2f46e588dab64d0fa13fd691a426405cb7b6a1"} Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383094 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-v5b57" event={"ID":"34b897ab-ce41-4cd5-bcae-cea3c0dea9e9","Type":"ContainerStarted","Data":"d1fcb80a9f83604cd9cc526d40f3aeec6026f7c8772347b1064ff90d72289853"} Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383428 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.383462 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.384759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xstlm" event={"ID":"f9398439-2570-42ec-b6fc-f9770b988c73","Type":"ContainerStarted","Data":"8afc6916ec9a9c7dd33fc43ed1b15de0c05e625f1672700c2abd0b3864a725b3"} Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.469473 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.923733 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-v5b57" podStartSLOduration=4.923709548 podStartE2EDuration="4.923709548s" podCreationTimestamp="2026-03-13 15:43:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:00.407113624 +0000 UTC m=+5939.306901353" watchObservedRunningTime="2026-03-13 15:44:00.923709548 +0000 UTC m=+5939.823497247" Mar 13 15:44:00 crc kubenswrapper[4907]: I0313 15:44:00.935827 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:44:00 crc kubenswrapper[4907]: W0313 15:44:00.948212 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04ca577b_ab8d_44a6_a598_1297bd5ab664.slice/crio-bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f WatchSource:0}: Error finding container bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f: Status 404 returned error can't find the container with id bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.050923 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.059898 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.070482 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-48f2-account-create-update-wvks8"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.079063 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gxjv5"] Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.396071 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-xstlm" event={"ID":"f9398439-2570-42ec-b6fc-f9770b988c73","Type":"ContainerStarted","Data":"b2340f62039b5ff0165b80754d58582bd36311f660eac8476fec48aa23ac78bf"} Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.397918 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" event={"ID":"04ca577b-ab8d-44a6-a598-1297bd5ab664","Type":"ContainerStarted","Data":"bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f"} Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.419908 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-xstlm" podStartSLOduration=2.419872174 podStartE2EDuration="2.419872174s" podCreationTimestamp="2026-03-13 15:43:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:01.410057406 +0000 UTC m=+5940.309845095" watchObservedRunningTime="2026-03-13 15:44:01.419872174 +0000 UTC m=+5940.319659883" Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.795361 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058" path="/var/lib/kubelet/pods/9c51cd5f-cfaf-4c1a-9aa2-e7ade1b79058/volumes" Mar 13 15:44:01 crc kubenswrapper[4907]: I0313 15:44:01.796808 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03a1328-cfc5-4ff6-88b9-6523f4378708" path="/var/lib/kubelet/pods/b03a1328-cfc5-4ff6-88b9-6523f4378708/volumes" Mar 13 15:44:02 crc kubenswrapper[4907]: I0313 15:44:02.406294 4907 generic.go:334] "Generic (PLEG): container finished" podID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerID="7157d1dc97b867b0493ef78bf3ed0dea2af629c4ae1bace22c4fa35387dfbe4f" exitCode=0 Mar 13 15:44:02 crc kubenswrapper[4907]: I0313 15:44:02.406381 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" event={"ID":"04ca577b-ab8d-44a6-a598-1297bd5ab664","Type":"ContainerDied","Data":"7157d1dc97b867b0493ef78bf3ed0dea2af629c4ae1bace22c4fa35387dfbe4f"} Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.702082 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.831151 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") pod \"04ca577b-ab8d-44a6-a598-1297bd5ab664\" (UID: \"04ca577b-ab8d-44a6-a598-1297bd5ab664\") " Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.836109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k" (OuterVolumeSpecName: "kube-api-access-76z5k") pod "04ca577b-ab8d-44a6-a598-1297bd5ab664" (UID: "04ca577b-ab8d-44a6-a598-1297bd5ab664"). InnerVolumeSpecName "kube-api-access-76z5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:03 crc kubenswrapper[4907]: I0313 15:44:03.933388 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76z5k\" (UniqueName: \"kubernetes.io/projected/04ca577b-ab8d-44a6-a598-1297bd5ab664-kube-api-access-76z5k\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.434608 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" event={"ID":"04ca577b-ab8d-44a6-a598-1297bd5ab664","Type":"ContainerDied","Data":"bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f"} Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.434658 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd9a2d3454311f65870318872731b5b1e0ae357a02e57d212f00b5e59db8956f" Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.434667 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556944-gcp6x" Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.758693 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.766045 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556938-ptksb"] Mar 13 15:44:04 crc kubenswrapper[4907]: I0313 15:44:04.783090 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:04 crc kubenswrapper[4907]: E0313 15:44:04.783484 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:05 crc kubenswrapper[4907]: I0313 15:44:05.795000 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5048c0-3add-450a-94c0-c068ceca72a4" path="/var/lib/kubelet/pods/6a5048c0-3add-450a-94c0-c068ceca72a4/volumes" Mar 13 15:44:07 crc kubenswrapper[4907]: I0313 15:44:07.029982 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:44:07 crc kubenswrapper[4907]: I0313 15:44:07.042097 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-4gtrw"] Mar 13 15:44:07 crc kubenswrapper[4907]: I0313 15:44:07.796872 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f23a73ee-57ff-4d58-8812-b72f624b7739" path="/var/lib/kubelet/pods/f23a73ee-57ff-4d58-8812-b72f624b7739/volumes" Mar 13 15:44:19 crc kubenswrapper[4907]: I0313 15:44:19.786511 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:19 crc kubenswrapper[4907]: E0313 15:44:19.787126 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:20 crc kubenswrapper[4907]: I0313 15:44:20.060103 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:44:20 crc kubenswrapper[4907]: I0313 15:44:20.067440 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mf86w"] Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.194858 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:44:21 crc kubenswrapper[4907]: E0313 15:44:21.195583 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerName="oc" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.195599 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerName="oc" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.195851 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" containerName="oc" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.196595 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.207231 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.279169 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.279292 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.381539 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.382017 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.382732 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.399504 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"octavia-db-create-ghczh\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.566216 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:21 crc kubenswrapper[4907]: I0313 15:44:21.795198 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0056ce52-b92e-4f2d-b3b0-a7dff01207ff" path="/var/lib/kubelet/pods/0056ce52-b92e-4f2d-b3b0-a7dff01207ff/volumes" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.048379 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:44:22 crc kubenswrapper[4907]: W0313 15:44:22.050494 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0aa6e501_c175_44be_af00_02b8aac76bab.slice/crio-df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f WatchSource:0}: Error finding container df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f: Status 404 returned error can't find the container with id df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.535098 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.536674 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.538271 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.559272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.601594 4907 generic.go:334] "Generic (PLEG): container finished" podID="0aa6e501-c175-44be-af00-02b8aac76bab" containerID="ef160b1245008062ab4f675d742f68b73acd696e5db3bbb49ff009cadb324f56" exitCode=0 Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.601638 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ghczh" event={"ID":"0aa6e501-c175-44be-af00-02b8aac76bab","Type":"ContainerDied","Data":"ef160b1245008062ab4f675d742f68b73acd696e5db3bbb49ff009cadb324f56"} Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.601664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ghczh" event={"ID":"0aa6e501-c175-44be-af00-02b8aac76bab","Type":"ContainerStarted","Data":"df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f"} Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.604681 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.604894 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.707201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.707282 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.708108 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.741089 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"octavia-6234-account-create-update-wjdx9\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:22 crc kubenswrapper[4907]: I0313 15:44:22.857255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.324936 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.614567 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerStarted","Data":"1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9"} Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.614935 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerStarted","Data":"e27a1243847151a896b96d511cbc60921c81660ade39e55cfd361a3a6ded5d7a"} Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.632208 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-6234-account-create-update-wjdx9" podStartSLOduration=1.632187283 podStartE2EDuration="1.632187283s" podCreationTimestamp="2026-03-13 15:44:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:23.629463598 +0000 UTC m=+5962.529251287" watchObservedRunningTime="2026-03-13 15:44:23.632187283 +0000 UTC m=+5962.531974982" Mar 13 15:44:23 crc kubenswrapper[4907]: I0313 15:44:23.952524 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.047476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") pod \"0aa6e501-c175-44be-af00-02b8aac76bab\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.047515 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") pod \"0aa6e501-c175-44be-af00-02b8aac76bab\" (UID: \"0aa6e501-c175-44be-af00-02b8aac76bab\") " Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.048233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0aa6e501-c175-44be-af00-02b8aac76bab" (UID: "0aa6e501-c175-44be-af00-02b8aac76bab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.053256 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2" (OuterVolumeSpecName: "kube-api-access-l5kw2") pod "0aa6e501-c175-44be-af00-02b8aac76bab" (UID: "0aa6e501-c175-44be-af00-02b8aac76bab"). InnerVolumeSpecName "kube-api-access-l5kw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.149336 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5kw2\" (UniqueName: \"kubernetes.io/projected/0aa6e501-c175-44be-af00-02b8aac76bab-kube-api-access-l5kw2\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.149374 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0aa6e501-c175-44be-af00-02b8aac76bab-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.622767 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-ghczh" event={"ID":"0aa6e501-c175-44be-af00-02b8aac76bab","Type":"ContainerDied","Data":"df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f"} Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.623147 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df9a2353d33e2427fe0c752dfa6b84b5852af4f10c0bc4157236881d9895340f" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.622816 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-ghczh" Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.625783 4907 generic.go:334] "Generic (PLEG): container finished" podID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerID="1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9" exitCode=0 Mar 13 15:44:24 crc kubenswrapper[4907]: I0313 15:44:24.625844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerDied","Data":"1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9"} Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.023847 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.085804 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") pod \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.085848 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") pod \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\" (UID: \"c420aeff-7a0b-4efa-b9c0-a93791eb0378\") " Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.086219 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c420aeff-7a0b-4efa-b9c0-a93791eb0378" (UID: "c420aeff-7a0b-4efa-b9c0-a93791eb0378"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.086396 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c420aeff-7a0b-4efa-b9c0-a93791eb0378-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.090630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b" (OuterVolumeSpecName: "kube-api-access-jgf6b") pod "c420aeff-7a0b-4efa-b9c0-a93791eb0378" (UID: "c420aeff-7a0b-4efa-b9c0-a93791eb0378"). InnerVolumeSpecName "kube-api-access-jgf6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.188302 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgf6b\" (UniqueName: \"kubernetes.io/projected/c420aeff-7a0b-4efa-b9c0-a93791eb0378-kube-api-access-jgf6b\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.644853 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-6234-account-create-update-wjdx9" event={"ID":"c420aeff-7a0b-4efa-b9c0-a93791eb0378","Type":"ContainerDied","Data":"e27a1243847151a896b96d511cbc60921c81660ade39e55cfd361a3a6ded5d7a"} Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.645170 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e27a1243847151a896b96d511cbc60921c81660ade39e55cfd361a3a6ded5d7a" Mar 13 15:44:26 crc kubenswrapper[4907]: I0313 15:44:26.644914 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-6234-account-create-update-wjdx9" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.225921 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:44:28 crc kubenswrapper[4907]: E0313 15:44:28.226311 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" containerName="mariadb-database-create" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226322 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" containerName="mariadb-database-create" Mar 13 15:44:28 crc kubenswrapper[4907]: E0313 15:44:28.226331 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerName="mariadb-account-create-update" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226337 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerName="mariadb-account-create-update" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226557 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" containerName="mariadb-database-create" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.226565 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" containerName="mariadb-account-create-update" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.227217 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.241792 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.328859 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.328958 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.430958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.431039 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.431829 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.448786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"octavia-persistence-db-create-p9hf2\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.546631 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.738018 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.739610 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.745801 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.747424 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.838481 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.838789 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.940298 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.940436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.941230 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:28 crc kubenswrapper[4907]: I0313 15:44:28.957588 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"octavia-1da8-account-create-update-82kpd\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.008796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.061056 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.305140 4907 scope.go:117] "RemoveContainer" containerID="e0b7f12342f112aa71fb3ae1454f11d3b6411a1575b99feb9bc8f40c4e53c905" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.354982 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.363800 4907 scope.go:117] "RemoveContainer" containerID="6f94031107f6bbf3f09d35d19fba9b0dfe08f6a055661f1e04241d246d42472d" Mar 13 15:44:29 crc kubenswrapper[4907]: W0313 15:44:29.369393 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b46488e_1fef_479f_94b6_1958e5174219.slice/crio-eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd WatchSource:0}: Error finding container eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd: Status 404 returned error can't find the container with id eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.388157 4907 scope.go:117] "RemoveContainer" containerID="d95988efdbf9f67c502e797aeeb6493599a1bcc1156251e60584683248ab36b0" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.441523 4907 scope.go:117] "RemoveContainer" containerID="19b5bd889c57ba8c1711b31659d515b279612f88bb0e0f645828fd40237e4402" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.542109 4907 scope.go:117] "RemoveContainer" containerID="a3a50fd2f676041cc1b8a25989baf35e7feed38466f29e552c042a8c33760d83" Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.677478 4907 generic.go:334] "Generic (PLEG): container finished" podID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerID="f56058903e847c31c47f45cadeb76ef300892a82a5c55a4fd28b47f0d9e1eeec" exitCode=0 Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.677547 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-p9hf2" event={"ID":"d218d11d-2afc-4f21-add2-af15cd8c50ad","Type":"ContainerDied","Data":"f56058903e847c31c47f45cadeb76ef300892a82a5c55a4fd28b47f0d9e1eeec"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.677572 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-p9hf2" event={"ID":"d218d11d-2afc-4f21-add2-af15cd8c50ad","Type":"ContainerStarted","Data":"22a8b1c8ecbde85926fedcdd711c6c938300b78064fcaf19ed68711cc58f703a"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.680135 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerStarted","Data":"e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.680161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerStarted","Data":"eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd"} Mar 13 15:44:29 crc kubenswrapper[4907]: I0313 15:44:29.727166 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-1da8-account-create-update-82kpd" podStartSLOduration=1.7271425969999998 podStartE2EDuration="1.727142597s" podCreationTimestamp="2026-03-13 15:44:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:44:29.716540887 +0000 UTC m=+5968.616328576" watchObservedRunningTime="2026-03-13 15:44:29.727142597 +0000 UTC m=+5968.626930286" Mar 13 15:44:30 crc kubenswrapper[4907]: I0313 15:44:30.690753 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b46488e-1fef-479f-94b6-1958e5174219" containerID="e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb" exitCode=0 Mar 13 15:44:30 crc kubenswrapper[4907]: I0313 15:44:30.690828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerDied","Data":"e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb"} Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.019511 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.182679 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") pod \"d218d11d-2afc-4f21-add2-af15cd8c50ad\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.182855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") pod \"d218d11d-2afc-4f21-add2-af15cd8c50ad\" (UID: \"d218d11d-2afc-4f21-add2-af15cd8c50ad\") " Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.183296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d218d11d-2afc-4f21-add2-af15cd8c50ad" (UID: "d218d11d-2afc-4f21-add2-af15cd8c50ad"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.183554 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d218d11d-2afc-4f21-add2-af15cd8c50ad-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.189360 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx" (OuterVolumeSpecName: "kube-api-access-jsfdx") pod "d218d11d-2afc-4f21-add2-af15cd8c50ad" (UID: "d218d11d-2afc-4f21-add2-af15cd8c50ad"). InnerVolumeSpecName "kube-api-access-jsfdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.285340 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsfdx\" (UniqueName: \"kubernetes.io/projected/d218d11d-2afc-4f21-add2-af15cd8c50ad-kube-api-access-jsfdx\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.701942 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-p9hf2" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.701937 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-p9hf2" event={"ID":"d218d11d-2afc-4f21-add2-af15cd8c50ad","Type":"ContainerDied","Data":"22a8b1c8ecbde85926fedcdd711c6c938300b78064fcaf19ed68711cc58f703a"} Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.702425 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22a8b1c8ecbde85926fedcdd711c6c938300b78064fcaf19ed68711cc58f703a" Mar 13 15:44:31 crc kubenswrapper[4907]: I0313 15:44:31.794386 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:31 crc kubenswrapper[4907]: E0313 15:44:31.794775 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.088008 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.115022 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") pod \"3b46488e-1fef-479f-94b6-1958e5174219\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.115243 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") pod \"3b46488e-1fef-479f-94b6-1958e5174219\" (UID: \"3b46488e-1fef-479f-94b6-1958e5174219\") " Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.115777 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3b46488e-1fef-479f-94b6-1958e5174219" (UID: "3b46488e-1fef-479f-94b6-1958e5174219"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.116189 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b46488e-1fef-479f-94b6-1958e5174219-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.120598 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg" (OuterVolumeSpecName: "kube-api-access-jz5hg") pod "3b46488e-1fef-479f-94b6-1958e5174219" (UID: "3b46488e-1fef-479f-94b6-1958e5174219"). InnerVolumeSpecName "kube-api-access-jz5hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.124385 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zb9k2" podUID="1cf43c88-f066-4979-8590-b8e6668aab09" containerName="ovn-controller" probeResult="failure" output=< Mar 13 15:44:32 crc kubenswrapper[4907]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 15:44:32 crc kubenswrapper[4907]: > Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.137546 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.161868 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-v5b57" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.219213 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz5hg\" (UniqueName: \"kubernetes.io/projected/3b46488e-1fef-479f-94b6-1958e5174219-kube-api-access-jz5hg\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.254525 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:32 crc kubenswrapper[4907]: E0313 15:44:32.255076 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b46488e-1fef-479f-94b6-1958e5174219" containerName="mariadb-account-create-update" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255095 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b46488e-1fef-479f-94b6-1958e5174219" containerName="mariadb-account-create-update" Mar 13 15:44:32 crc kubenswrapper[4907]: E0313 15:44:32.255124 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerName="mariadb-database-create" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255132 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerName="mariadb-database-create" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255417 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" containerName="mariadb-database-create" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.255437 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b46488e-1fef-479f-94b6-1958e5174219" containerName="mariadb-account-create-update" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.256407 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.258538 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.271530 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320403 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320530 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.320617 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422201 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422409 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422442 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.422775 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.423798 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.423858 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.423911 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.425560 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.455695 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ovn-controller-zb9k2-config-4n289\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.583134 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.740826 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-1da8-account-create-update-82kpd" event={"ID":"3b46488e-1fef-479f-94b6-1958e5174219","Type":"ContainerDied","Data":"eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd"} Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.741190 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeb8ce48309afbf39df362ceed325c415fd044fc1e43aef2b30ca4f7acab66dd" Mar 13 15:44:32 crc kubenswrapper[4907]: I0313 15:44:32.740864 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-1da8-account-create-update-82kpd" Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.101517 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:33 crc kubenswrapper[4907]: W0313 15:44:33.102898 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecd299e1_64f9_4ad1_9ff8_af388e993db9.slice/crio-c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437 WatchSource:0}: Error finding container c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437: Status 404 returned error can't find the container with id c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437 Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.751697 4907 generic.go:334] "Generic (PLEG): container finished" podID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerID="219cbe1415b39a75ce8171bb61ada0538ae87eb0c8b470a088b03fdc9289f0b1" exitCode=0 Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.751763 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-4n289" event={"ID":"ecd299e1-64f9-4ad1-9ff8-af388e993db9","Type":"ContainerDied","Data":"219cbe1415b39a75ce8171bb61ada0538ae87eb0c8b470a088b03fdc9289f0b1"} Mar 13 15:44:33 crc kubenswrapper[4907]: I0313 15:44:33.752042 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-4n289" event={"ID":"ecd299e1-64f9-4ad1-9ff8-af388e993db9","Type":"ContainerStarted","Data":"c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437"} Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.165068 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-6db876fcbd-4wtkj"] Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.168738 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.171873 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.172080 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-m6wbw" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.178310 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.180171 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-6db876fcbd-4wtkj"] Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.258847 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data-merged\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259009 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-scripts\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259045 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-octavia-run\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.259399 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-combined-ca-bundle\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.360933 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-scripts\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.360983 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361088 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-octavia-run\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361126 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-combined-ca-bundle\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361240 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data-merged\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-octavia-run\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.361780 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data-merged\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.367325 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-combined-ca-bundle\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.367614 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-config-data\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.371601 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c868ec4-5e49-4be1-a1eb-c4748753aecb-scripts\") pod \"octavia-api-6db876fcbd-4wtkj\" (UID: \"1c868ec4-5e49-4be1-a1eb-c4748753aecb\") " pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:34 crc kubenswrapper[4907]: I0313 15:44:34.486983 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.137093 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-6db876fcbd-4wtkj"] Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.174233 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285103 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285492 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285613 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.285737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.286023 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.286275 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") pod \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\" (UID: \"ecd299e1-64f9-4ad1-9ff8-af388e993db9\") " Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.286995 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run" (OuterVolumeSpecName: "var-run") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287046 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287076 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts" (OuterVolumeSpecName: "scripts") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.287625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.293966 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj" (OuterVolumeSpecName: "kube-api-access-sf9zj") pod "ecd299e1-64f9-4ad1-9ff8-af388e993db9" (UID: "ecd299e1-64f9-4ad1-9ff8-af388e993db9"). InnerVolumeSpecName "kube-api-access-sf9zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388404 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388663 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388723 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388774 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ecd299e1-64f9-4ad1-9ff8-af388e993db9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388837 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf9zj\" (UniqueName: \"kubernetes.io/projected/ecd299e1-64f9-4ad1-9ff8-af388e993db9-kube-api-access-sf9zj\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.388937 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/ecd299e1-64f9-4ad1-9ff8-af388e993db9-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.772514 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"86b85bba126122ac436f8e2f67651035d64e2de892225b44d780ad2018449177"} Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.773969 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-4n289" event={"ID":"ecd299e1-64f9-4ad1-9ff8-af388e993db9","Type":"ContainerDied","Data":"c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437"} Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.773996 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4a92902f8ae073c889d6b0292d8ab146828a907c2ec89705db070f340ffd437" Mar 13 15:44:35 crc kubenswrapper[4907]: I0313 15:44:35.774037 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-4n289" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.248744 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.257663 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zb9k2-config-4n289"] Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.294473 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:36 crc kubenswrapper[4907]: E0313 15:44:36.294916 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerName="ovn-config" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.294932 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerName="ovn-config" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.295118 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" containerName="ovn-config" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.295775 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.299210 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.309434 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.409974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410108 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410162 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410191 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410410 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.410642 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512706 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512800 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512913 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.512977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.513132 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.513156 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.513186 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.514097 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.515990 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.531897 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"ovn-controller-zb9k2-config-jvnzd\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:36 crc kubenswrapper[4907]: I0313 15:44:36.622487 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.086351 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:37 crc kubenswrapper[4907]: W0313 15:44:37.092064 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod405d2caf_b667_4833_9e03_f6eaaff9f8c1.slice/crio-69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369 WatchSource:0}: Error finding container 69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369: Status 404 returned error can't find the container with id 69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369 Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.125578 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zb9k2" Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.799528 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd299e1-64f9-4ad1-9ff8-af388e993db9" path="/var/lib/kubelet/pods/ecd299e1-64f9-4ad1-9ff8-af388e993db9/volumes" Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.805560 4907 generic.go:334] "Generic (PLEG): container finished" podID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerID="fea279a9a11b2a38768139c4f75fb40e14428aafd4173f88229129062a569479" exitCode=0 Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.805629 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-jvnzd" event={"ID":"405d2caf-b667-4833-9e03-f6eaaff9f8c1","Type":"ContainerDied","Data":"fea279a9a11b2a38768139c4f75fb40e14428aafd4173f88229129062a569479"} Mar 13 15:44:37 crc kubenswrapper[4907]: I0313 15:44:37.805686 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-jvnzd" event={"ID":"405d2caf-b667-4833-9e03-f6eaaff9f8c1","Type":"ContainerStarted","Data":"69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369"} Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.559804 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575700 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575646 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575892 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575926 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.575980 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576069 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") pod \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\" (UID: \"405d2caf-b667-4833-9e03-f6eaaff9f8c1\") " Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576464 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576505 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.576532 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run" (OuterVolumeSpecName: "var-run") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.578110 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.579062 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts" (OuterVolumeSpecName: "scripts") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.582070 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r" (OuterVolumeSpecName: "kube-api-access-x4w5r") pod "405d2caf-b667-4833-9e03-f6eaaff9f8c1" (UID: "405d2caf-b667-4833-9e03-f6eaaff9f8c1"). InnerVolumeSpecName "kube-api-access-x4w5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677692 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677936 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/405d2caf-b667-4833-9e03-f6eaaff9f8c1-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677949 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4w5r\" (UniqueName: \"kubernetes.io/projected/405d2caf-b667-4833-9e03-f6eaaff9f8c1-kube-api-access-x4w5r\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677959 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.677969 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/405d2caf-b667-4833-9e03-f6eaaff9f8c1-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.873637 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-jvnzd" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.873782 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-jvnzd" event={"ID":"405d2caf-b667-4833-9e03-f6eaaff9f8c1","Type":"ContainerDied","Data":"69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369"} Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.873819 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69e762bd095eb220585503bedb5a31de1a473c47dfddf7b51f8a8e391d564369" Mar 13 15:44:44 crc kubenswrapper[4907]: I0313 15:44:44.875299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"55fb7423f1d00e837760507cfb3fc4502b1b604d521e1f69652f3853242bd445"} Mar 13 15:44:45 crc kubenswrapper[4907]: I0313 15:44:45.638521 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.193791 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zb9k2-config-jvnzd"] Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.196935 4907 generic.go:334] "Generic (PLEG): container finished" podID="1c868ec4-5e49-4be1-a1eb-c4748753aecb" containerID="55fb7423f1d00e837760507cfb3fc4502b1b604d521e1f69652f3853242bd445" exitCode=0 Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.196966 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerDied","Data":"55fb7423f1d00e837760507cfb3fc4502b1b604d521e1f69652f3853242bd445"} Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.199546 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:46 crc kubenswrapper[4907]: E0313 15:44:46.200239 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerName="ovn-config" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.200276 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerName="ovn-config" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.200554 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" containerName="ovn-config" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.201453 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.205131 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.210983 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337678 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337749 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337779 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337831 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.337907 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.439927 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.439976 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440026 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440047 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440112 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440504 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440656 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.440977 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.442642 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.442735 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.461486 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"ovn-controller-zb9k2-config-2d8dv\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.695254 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:46 crc kubenswrapper[4907]: I0313 15:44:46.783280 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:46 crc kubenswrapper[4907]: E0313 15:44:46.783641 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.185570 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:47 crc kubenswrapper[4907]: W0313 15:44:47.192849 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96a87d0f_b3d7_447b_a8db_1ad11b93f37a.slice/crio-e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768 WatchSource:0}: Error finding container e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768: Status 404 returned error can't find the container with id e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768 Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.209208 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-2d8dv" event={"ID":"96a87d0f-b3d7-447b-a8db-1ad11b93f37a","Type":"ContainerStarted","Data":"e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768"} Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.211311 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"432213e092050825f06d1f9ead3214f9559b8e4a02723df2cf0dd05d48859ce9"} Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.211345 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-6db876fcbd-4wtkj" event={"ID":"1c868ec4-5e49-4be1-a1eb-c4748753aecb","Type":"ContainerStarted","Data":"1a0a2f6ef0a9cb2e75b38a9ad6ca30d1ad8381dc759967de44f409f3e35e7c40"} Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.211594 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.250704 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-6db876fcbd-4wtkj" podStartSLOduration=3.88665379 podStartE2EDuration="13.250685772s" podCreationTimestamp="2026-03-13 15:44:34 +0000 UTC" firstStartedPulling="2026-03-13 15:44:35.168094025 +0000 UTC m=+5974.067881714" lastFinishedPulling="2026-03-13 15:44:44.532126007 +0000 UTC m=+5983.431913696" observedRunningTime="2026-03-13 15:44:47.243152357 +0000 UTC m=+5986.142940056" watchObservedRunningTime="2026-03-13 15:44:47.250685772 +0000 UTC m=+5986.150473461" Mar 13 15:44:47 crc kubenswrapper[4907]: I0313 15:44:47.792827 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405d2caf-b667-4833-9e03-f6eaaff9f8c1" path="/var/lib/kubelet/pods/405d2caf-b667-4833-9e03-f6eaaff9f8c1/volumes" Mar 13 15:44:48 crc kubenswrapper[4907]: I0313 15:44:48.222649 4907 generic.go:334] "Generic (PLEG): container finished" podID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerID="37e65533c5fc864e730a3583c9b7118a1665874961e44f735e26e376e6b5f39d" exitCode=0 Mar 13 15:44:48 crc kubenswrapper[4907]: I0313 15:44:48.223027 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-2d8dv" event={"ID":"96a87d0f-b3d7-447b-a8db-1ad11b93f37a","Type":"ContainerDied","Data":"37e65533c5fc864e730a3583c9b7118a1665874961e44f735e26e376e6b5f39d"} Mar 13 15:44:48 crc kubenswrapper[4907]: I0313 15:44:48.223153 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.664387 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805139 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805267 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805349 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805380 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805468 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") pod \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\" (UID: \"96a87d0f-b3d7-447b-a8db-1ad11b93f37a\") " Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805611 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run" (OuterVolumeSpecName: "var-run") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.805643 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.806015 4907 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.806037 4907 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.806048 4907 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.818694 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc" (OuterVolumeSpecName: "kube-api-access-d8zkc") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "kube-api-access-d8zkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.856498 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.856711 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts" (OuterVolumeSpecName: "scripts") pod "96a87d0f-b3d7-447b-a8db-1ad11b93f37a" (UID: "96a87d0f-b3d7-447b-a8db-1ad11b93f37a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.908175 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.908217 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8zkc\" (UniqueName: \"kubernetes.io/projected/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-kube-api-access-d8zkc\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:49 crc kubenswrapper[4907]: I0313 15:44:49.908231 4907 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/96a87d0f-b3d7-447b-a8db-1ad11b93f37a-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.260472 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zb9k2-config-2d8dv" event={"ID":"96a87d0f-b3d7-447b-a8db-1ad11b93f37a","Type":"ContainerDied","Data":"e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768"} Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.260550 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01fd45fb886de621cf61d6ee0706af4f41676bc2037299d72cd797bdbf65768" Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.260636 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zb9k2-config-2d8dv" Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.759613 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:50 crc kubenswrapper[4907]: I0313 15:44:50.777551 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zb9k2-config-2d8dv"] Mar 13 15:44:51 crc kubenswrapper[4907]: I0313 15:44:51.792459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" path="/var/lib/kubelet/pods/96a87d0f-b3d7-447b-a8db-1ad11b93f37a/volumes" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.646461 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:57 crc kubenswrapper[4907]: E0313 15:44:57.647431 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerName="ovn-config" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.647450 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerName="ovn-config" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.647707 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a87d0f-b3d7-447b-a8db-1ad11b93f37a" containerName="ovn-config" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.648687 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.657154 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.657275 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.657606 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.674168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.750662 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4df4068e-249f-47ea-a84a-3c178c27bdf3-hm-ports\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.751179 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-scripts\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.751242 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data-merged\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.751486 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.782852 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:44:57 crc kubenswrapper[4907]: E0313 15:44:57.783150 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854033 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854404 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4df4068e-249f-47ea-a84a-3c178c27bdf3-hm-ports\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854660 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-scripts\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.854745 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data-merged\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.855440 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data-merged\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.856994 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/4df4068e-249f-47ea-a84a-3c178c27bdf3-hm-ports\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.866459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-config-data\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.866922 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4df4068e-249f-47ea-a84a-3c178c27bdf3-scripts\") pod \"octavia-rsyslog-zbs6z\" (UID: \"4df4068e-249f-47ea-a84a-3c178c27bdf3\") " pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:57 crc kubenswrapper[4907]: I0313 15:44:57.973414 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.596518 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:58 crc kubenswrapper[4907]: W0313 15:44:58.597073 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4df4068e_249f_47ea_a84a_3c178c27bdf3.slice/crio-0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5 WatchSource:0}: Error finding container 0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5: Status 404 returned error can't find the container with id 0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5 Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.689209 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.691002 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.697871 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.712947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.723588 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-zbs6z"] Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.775439 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.775742 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.878764 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.879241 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.880525 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:58 crc kubenswrapper[4907]: I0313 15:44:58.887339 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"octavia-image-upload-6f45c4fb85-gvk5g\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:59 crc kubenswrapper[4907]: I0313 15:44:59.037228 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:44:59 crc kubenswrapper[4907]: I0313 15:44:59.361302 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerStarted","Data":"0d577fd9519d47b2c397ce97d78c1842b141c20b3852e2b82b1bddfd1f827fc5"} Mar 13 15:44:59 crc kubenswrapper[4907]: I0313 15:44:59.568144 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.140196 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.142308 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.145186 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.145388 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.163987 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.203201 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.203263 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.203290 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.305219 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.305273 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.305296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.306972 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.314500 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.325819 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"collect-profiles-29556945-vp2lp\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.391712 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerStarted","Data":"779798f2bbcef27c1f90b8ec15e35048f6beabc9118118529e9bc26eeed24e32"} Mar 13 15:45:00 crc kubenswrapper[4907]: I0313 15:45:00.466977 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:01 crc kubenswrapper[4907]: I0313 15:45:01.022110 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 15:45:01 crc kubenswrapper[4907]: I0313 15:45:01.412685 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerStarted","Data":"c19a72a707aaaf1cb3cc867c2a276800bcfd60b1ace8c673b72611d9326d6913"} Mar 13 15:45:02 crc kubenswrapper[4907]: I0313 15:45:02.424999 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" event={"ID":"329cddab-6958-4dbb-a2d0-baef50885413","Type":"ContainerStarted","Data":"114a7e04a442a61d9b64fe621624aaf4b09fa81b5a1bf440bf909e958cdc0ffe"} Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.439673 4907 generic.go:334] "Generic (PLEG): container finished" podID="4df4068e-249f-47ea-a84a-3c178c27bdf3" containerID="c19a72a707aaaf1cb3cc867c2a276800bcfd60b1ace8c673b72611d9326d6913" exitCode=0 Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.439859 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerDied","Data":"c19a72a707aaaf1cb3cc867c2a276800bcfd60b1ace8c673b72611d9326d6913"} Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.449431 4907 generic.go:334] "Generic (PLEG): container finished" podID="329cddab-6958-4dbb-a2d0-baef50885413" containerID="65a8dfeb436104f422bc6ebeae3f1c135d2d01201103445d064625f2317603ad" exitCode=0 Mar 13 15:45:03 crc kubenswrapper[4907]: I0313 15:45:03.449506 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" event={"ID":"329cddab-6958-4dbb-a2d0-baef50885413","Type":"ContainerDied","Data":"65a8dfeb436104f422bc6ebeae3f1c135d2d01201103445d064625f2317603ad"} Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.409452 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.491051 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.491527 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp" event={"ID":"329cddab-6958-4dbb-a2d0-baef50885413","Type":"ContainerDied","Data":"114a7e04a442a61d9b64fe621624aaf4b09fa81b5a1bf440bf909e958cdc0ffe"} Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.491552 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="114a7e04a442a61d9b64fe621624aaf4b09fa81b5a1bf440bf909e958cdc0ffe" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.550658 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") pod \"329cddab-6958-4dbb-a2d0-baef50885413\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.550782 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") pod \"329cddab-6958-4dbb-a2d0-baef50885413\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.550808 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") pod \"329cddab-6958-4dbb-a2d0-baef50885413\" (UID: \"329cddab-6958-4dbb-a2d0-baef50885413\") " Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.553620 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume" (OuterVolumeSpecName: "config-volume") pod "329cddab-6958-4dbb-a2d0-baef50885413" (UID: "329cddab-6958-4dbb-a2d0-baef50885413"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.562174 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk" (OuterVolumeSpecName: "kube-api-access-rbxfk") pod "329cddab-6958-4dbb-a2d0-baef50885413" (UID: "329cddab-6958-4dbb-a2d0-baef50885413"). InnerVolumeSpecName "kube-api-access-rbxfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.572087 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "329cddab-6958-4dbb-a2d0-baef50885413" (UID: "329cddab-6958-4dbb-a2d0-baef50885413"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.652709 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/329cddab-6958-4dbb-a2d0-baef50885413-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.652742 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbxfk\" (UniqueName: \"kubernetes.io/projected/329cddab-6958-4dbb-a2d0-baef50885413-kube-api-access-rbxfk\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:05 crc kubenswrapper[4907]: I0313 15:45:05.652753 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/329cddab-6958-4dbb-a2d0-baef50885413-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.181148 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:45:06 crc kubenswrapper[4907]: E0313 15:45:06.182052 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="329cddab-6958-4dbb-a2d0-baef50885413" containerName="collect-profiles" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.182078 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="329cddab-6958-4dbb-a2d0-baef50885413" containerName="collect-profiles" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.182344 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="329cddab-6958-4dbb-a2d0-baef50885413" containerName="collect-profiles" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.183736 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.188168 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.194828 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.264740 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.265072 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.265171 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.265342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366342 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366392 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366486 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.366525 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.367115 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.374191 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.376608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.380739 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"octavia-db-sync-rjfnd\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.505741 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.512579 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-zbs6z" event={"ID":"4df4068e-249f-47ea-a84a-3c178c27bdf3","Type":"ContainerStarted","Data":"74a44fccf52afbb9c6ff10cceccc82caa3010599c5105999945e23c4bc3970d5"} Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.512837 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.514805 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556900-cxbzm"] Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.532366 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:06 crc kubenswrapper[4907]: I0313 15:45:06.536398 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-zbs6z" podStartSLOduration=2.909128543 podStartE2EDuration="9.53637243s" podCreationTimestamp="2026-03-13 15:44:57 +0000 UTC" firstStartedPulling="2026-03-13 15:44:58.599486781 +0000 UTC m=+5997.499274470" lastFinishedPulling="2026-03-13 15:45:05.226730668 +0000 UTC m=+6004.126518357" observedRunningTime="2026-03-13 15:45:06.530014457 +0000 UTC m=+6005.429802156" watchObservedRunningTime="2026-03-13 15:45:06.53637243 +0000 UTC m=+6005.436160129" Mar 13 15:45:07 crc kubenswrapper[4907]: I0313 15:45:07.874250 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48ae7008-a336-47d8-a3da-fd198db91cba" path="/var/lib/kubelet/pods/48ae7008-a336-47d8-a3da-fd198db91cba/volumes" Mar 13 15:45:08 crc kubenswrapper[4907]: I0313 15:45:08.783203 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:08 crc kubenswrapper[4907]: E0313 15:45:08.783767 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:09 crc kubenswrapper[4907]: I0313 15:45:09.448748 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:45:09 crc kubenswrapper[4907]: I0313 15:45:09.952180 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-6db876fcbd-4wtkj" Mar 13 15:45:10 crc kubenswrapper[4907]: I0313 15:45:10.520407 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:45:10 crc kubenswrapper[4907]: I0313 15:45:10.579693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerStarted","Data":"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc"} Mar 13 15:45:10 crc kubenswrapper[4907]: I0313 15:45:10.582699 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerStarted","Data":"c0ee84ab0839ac1ec5f8ef593950c63bef20d37354985ce33bc2c5c08ed061a9"} Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.591546 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" exitCode=0 Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.591723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerDied","Data":"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc"} Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.594721 4907 generic.go:334] "Generic (PLEG): container finished" podID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerID="94c59567d9e99f3b6037554dff6847cedf9844e341b49a4a37077ea52d3428be" exitCode=0 Mar 13 15:45:11 crc kubenswrapper[4907]: I0313 15:45:11.594772 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerDied","Data":"94c59567d9e99f3b6037554dff6847cedf9844e341b49a4a37077ea52d3428be"} Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.604519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerStarted","Data":"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1"} Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.607225 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerStarted","Data":"4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663"} Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.627595 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" podStartSLOduration=4.037097773 podStartE2EDuration="14.627569314s" podCreationTimestamp="2026-03-13 15:44:58 +0000 UTC" firstStartedPulling="2026-03-13 15:44:59.579802235 +0000 UTC m=+5998.479589914" lastFinishedPulling="2026-03-13 15:45:10.170273766 +0000 UTC m=+6009.070061455" observedRunningTime="2026-03-13 15:45:12.616783659 +0000 UTC m=+6011.516571348" watchObservedRunningTime="2026-03-13 15:45:12.627569314 +0000 UTC m=+6011.527356993" Mar 13 15:45:12 crc kubenswrapper[4907]: I0313 15:45:12.633731 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-rjfnd" podStartSLOduration=6.633712071 podStartE2EDuration="6.633712071s" podCreationTimestamp="2026-03-13 15:45:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:45:12.632037805 +0000 UTC m=+6011.531825494" watchObservedRunningTime="2026-03-13 15:45:12.633712071 +0000 UTC m=+6011.533499760" Mar 13 15:45:13 crc kubenswrapper[4907]: I0313 15:45:13.001398 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-zbs6z" Mar 13 15:45:14 crc kubenswrapper[4907]: I0313 15:45:14.626610 4907 generic.go:334] "Generic (PLEG): container finished" podID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerID="4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663" exitCode=0 Mar 13 15:45:14 crc kubenswrapper[4907]: I0313 15:45:14.626693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerDied","Data":"4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663"} Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.010382 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.187342 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.188179 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.188308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.188500 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") pod \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\" (UID: \"aefbf41f-fa5a-4763-af85-3c9e8df33b6a\") " Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.192279 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts" (OuterVolumeSpecName: "scripts") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.193561 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data" (OuterVolumeSpecName: "config-data") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.218211 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.220382 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aefbf41f-fa5a-4763-af85-3c9e8df33b6a" (UID: "aefbf41f-fa5a-4763-af85-3c9e8df33b6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291017 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291051 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291065 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.291080 4907 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/aefbf41f-fa5a-4763-af85-3c9e8df33b6a-config-data-merged\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.651298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-rjfnd" event={"ID":"aefbf41f-fa5a-4763-af85-3c9e8df33b6a","Type":"ContainerDied","Data":"c0ee84ab0839ac1ec5f8ef593950c63bef20d37354985ce33bc2c5c08ed061a9"} Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.651343 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0ee84ab0839ac1ec5f8ef593950c63bef20d37354985ce33bc2c5c08ed061a9" Mar 13 15:45:16 crc kubenswrapper[4907]: I0313 15:45:16.651385 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-rjfnd" Mar 13 15:45:20 crc kubenswrapper[4907]: I0313 15:45:20.782436 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:20 crc kubenswrapper[4907]: E0313 15:45:20.784343 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:29 crc kubenswrapper[4907]: I0313 15:45:29.666216 4907 scope.go:117] "RemoveContainer" containerID="189d9f0966abd854ff0ae3d577047f2f411f08ff9ec8a2c05b10583fd1185431" Mar 13 15:45:33 crc kubenswrapper[4907]: I0313 15:45:33.784452 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:33 crc kubenswrapper[4907]: E0313 15:45:33.785257 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:38 crc kubenswrapper[4907]: I0313 15:45:38.806060 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:38 crc kubenswrapper[4907]: I0313 15:45:38.806798 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" containerID="cri-o://8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" gracePeriod=30 Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.334739 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.453385 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") pod \"7a6ac004-9828-4581-9147-88c96b1f5ef4\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.453923 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") pod \"7a6ac004-9828-4581-9147-88c96b1f5ef4\" (UID: \"7a6ac004-9828-4581-9147-88c96b1f5ef4\") " Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.496830 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7a6ac004-9828-4581-9147-88c96b1f5ef4" (UID: "7a6ac004-9828-4581-9147-88c96b1f5ef4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.509849 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "7a6ac004-9828-4581-9147-88c96b1f5ef4" (UID: "7a6ac004-9828-4581-9147-88c96b1f5ef4"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.556469 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7a6ac004-9828-4581-9147-88c96b1f5ef4-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.556516 4907 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/7a6ac004-9828-4581-9147-88c96b1f5ef4-amphora-image\") on node \"crc\" DevicePath \"\"" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847179 4907 generic.go:334] "Generic (PLEG): container finished" podID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" exitCode=0 Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerDied","Data":"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1"} Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" event={"ID":"7a6ac004-9828-4581-9147-88c96b1f5ef4","Type":"ContainerDied","Data":"779798f2bbcef27c1f90b8ec15e35048f6beabc9118118529e9bc26eeed24e32"} Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847280 4907 scope.go:117] "RemoveContainer" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.847424 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-6f45c4fb85-gvk5g" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.871543 4907 scope.go:117] "RemoveContainer" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.876089 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.889529 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-6f45c4fb85-gvk5g"] Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.893389 4907 scope.go:117] "RemoveContainer" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" Mar 13 15:45:39 crc kubenswrapper[4907]: E0313 15:45:39.895341 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1\": container with ID starting with 8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1 not found: ID does not exist" containerID="8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.895369 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1"} err="failed to get container status \"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1\": rpc error: code = NotFound desc = could not find container \"8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1\": container with ID starting with 8c362fd432faeca33e638f398523ea974dc977713306f90c2ab5a670b92e3cb1 not found: ID does not exist" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.895388 4907 scope.go:117] "RemoveContainer" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" Mar 13 15:45:39 crc kubenswrapper[4907]: E0313 15:45:39.895697 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc\": container with ID starting with 74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc not found: ID does not exist" containerID="74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc" Mar 13 15:45:39 crc kubenswrapper[4907]: I0313 15:45:39.895737 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc"} err="failed to get container status \"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc\": rpc error: code = NotFound desc = could not find container \"74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc\": container with ID starting with 74f45f2ca03eeba83bae0697e69c57f95344fad1a6ad11c8a7d301ea69e6bdbc not found: ID does not exist" Mar 13 15:45:41 crc kubenswrapper[4907]: I0313 15:45:41.792403 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" path="/var/lib/kubelet/pods/7a6ac004-9828-4581-9147-88c96b1f5ef4/volumes" Mar 13 15:45:46 crc kubenswrapper[4907]: I0313 15:45:46.782971 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:46 crc kubenswrapper[4907]: E0313 15:45:46.783695 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:45:58 crc kubenswrapper[4907]: I0313 15:45:58.782677 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:45:58 crc kubenswrapper[4907]: E0313 15:45:58.783404 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.143809 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144537 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144550 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144557 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="octavia-db-sync" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144563 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="octavia-db-sync" Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144577 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144583 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="init" Mar 13 15:46:00 crc kubenswrapper[4907]: E0313 15:46:00.144608 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144615 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144827 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" containerName="octavia-db-sync" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.144852 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6ac004-9828-4581-9147-88c96b1f5ef4" containerName="octavia-amphora-httpd" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.145485 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.148178 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.148317 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.151221 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.159468 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.283026 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"auto-csr-approver-29556946-78rqg\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.384648 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"auto-csr-approver-29556946-78rqg\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.412720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"auto-csr-approver-29556946-78rqg\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.464514 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:00 crc kubenswrapper[4907]: I0313 15:46:00.923545 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:46:01 crc kubenswrapper[4907]: I0313 15:46:01.031924 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556946-78rqg" event={"ID":"2f0d1531-1d25-4331-a690-1fcbcb504071","Type":"ContainerStarted","Data":"9e7cadbe55b65797301703c241dfad204c1d7276015b0c5406cafc5fc5c850e2"} Mar 13 15:46:02 crc kubenswrapper[4907]: E0313 15:46:02.497075 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f0d1531_1d25_4331_a690_1fcbcb504071.slice/crio-conmon-971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.563717 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.566458 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.568272 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.568345 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.568669 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.580527 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737657 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-amphora-certs\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737706 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/57610eac-b9eb-4628-bd99-b3ebb35b2772-hm-ports\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737795 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-combined-ca-bundle\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737913 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-scripts\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737938 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.737986 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data-merged\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839643 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-amphora-certs\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839702 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/57610eac-b9eb-4628-bd99-b3ebb35b2772-hm-ports\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839760 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-combined-ca-bundle\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839796 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-scripts\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839818 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.839859 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data-merged\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.840564 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data-merged\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.841109 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/57610eac-b9eb-4628-bd99-b3ebb35b2772-hm-ports\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.846032 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-config-data\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.846296 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-scripts\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.860683 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-combined-ca-bundle\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.863785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/57610eac-b9eb-4628-bd99-b3ebb35b2772-amphora-certs\") pod \"octavia-healthmanager-hbbw4\" (UID: \"57610eac-b9eb-4628-bd99-b3ebb35b2772\") " pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:02 crc kubenswrapper[4907]: I0313 15:46:02.885002 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:03 crc kubenswrapper[4907]: I0313 15:46:03.052854 4907 generic.go:334] "Generic (PLEG): container finished" podID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerID="971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380" exitCode=0 Mar 13 15:46:03 crc kubenswrapper[4907]: I0313 15:46:03.053044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556946-78rqg" event={"ID":"2f0d1531-1d25-4331-a690-1fcbcb504071","Type":"ContainerDied","Data":"971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380"} Mar 13 15:46:03 crc kubenswrapper[4907]: I0313 15:46:03.602994 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.063907 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerStarted","Data":"27749e579eeb959a64f9197a7832237eb739e4daba11766edbfeb3f7aaf9ad37"} Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.064218 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerStarted","Data":"b035a1861472eee7e65e830e7bb53c1fa16793b9e742ac953f22caab1ad82095"} Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.231284 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-p69c5"] Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.233230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.236975 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.237287 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.245078 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-p69c5"] Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374164 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab5868a8-f083-4417-b7c0-f01dad538927-hm-ports\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374217 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-amphora-certs\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-scripts\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-combined-ca-bundle\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374357 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-config-data\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.374417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab5868a8-f083-4417-b7c0-f01dad538927-config-data-merged\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.425791 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-scripts\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475713 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-combined-ca-bundle\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475744 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-config-data\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475812 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab5868a8-f083-4417-b7c0-f01dad538927-config-data-merged\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475847 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab5868a8-f083-4417-b7c0-f01dad538927-hm-ports\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.475873 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-amphora-certs\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.476817 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/ab5868a8-f083-4417-b7c0-f01dad538927-config-data-merged\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.477953 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/ab5868a8-f083-4417-b7c0-f01dad538927-hm-ports\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.482025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-combined-ca-bundle\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.482248 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-scripts\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.485923 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-config-data\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.486597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/ab5868a8-f083-4417-b7c0-f01dad538927-amphora-certs\") pod \"octavia-housekeeping-p69c5\" (UID: \"ab5868a8-f083-4417-b7c0-f01dad538927\") " pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.558082 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.589187 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") pod \"2f0d1531-1d25-4331-a690-1fcbcb504071\" (UID: \"2f0d1531-1d25-4331-a690-1fcbcb504071\") " Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.592334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz" (OuterVolumeSpecName: "kube-api-access-lqghz") pod "2f0d1531-1d25-4331-a690-1fcbcb504071" (UID: "2f0d1531-1d25-4331-a690-1fcbcb504071"). InnerVolumeSpecName "kube-api-access-lqghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:04 crc kubenswrapper[4907]: I0313 15:46:04.691853 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqghz\" (UniqueName: \"kubernetes.io/projected/2f0d1531-1d25-4331-a690-1fcbcb504071-kube-api-access-lqghz\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.010870 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-7pld2"] Mar 13 15:46:05 crc kubenswrapper[4907]: E0313 15:46:05.011943 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerName="oc" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.012019 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerName="oc" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.012292 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" containerName="oc" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.022154 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.032420 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.032665 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.050567 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-7pld2"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.086166 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556946-78rqg" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.098001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556946-78rqg" event={"ID":"2f0d1531-1d25-4331-a690-1fcbcb504071","Type":"ContainerDied","Data":"9e7cadbe55b65797301703c241dfad204c1d7276015b0c5406cafc5fc5c850e2"} Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.098067 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e7cadbe55b65797301703c241dfad204c1d7276015b0c5406cafc5fc5c850e2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.207947 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f6c0a640-5177-4188-abc3-d934f75624bb-config-data-merged\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208019 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-amphora-certs\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208109 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f6c0a640-5177-4188-abc3-d934f75624bb-hm-ports\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208155 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-scripts\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-config-data\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.208211 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-combined-ca-bundle\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.254020 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-p69c5"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312205 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-scripts\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312302 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-config-data\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312332 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-combined-ca-bundle\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312382 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f6c0a640-5177-4188-abc3-d934f75624bb-config-data-merged\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312434 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-amphora-certs\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.312547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f6c0a640-5177-4188-abc3-d934f75624bb-hm-ports\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.313840 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f6c0a640-5177-4188-abc3-d934f75624bb-hm-ports\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.314242 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f6c0a640-5177-4188-abc3-d934f75624bb-config-data-merged\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.324312 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-combined-ca-bundle\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.329057 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-scripts\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.333621 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-amphora-certs\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.342308 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6c0a640-5177-4188-abc3-d934f75624bb-config-data\") pod \"octavia-worker-7pld2\" (UID: \"f6c0a640-5177-4188-abc3-d934f75624bb\") " pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.351508 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.509509 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.522224 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556940-5rmt4"] Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.793262 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d20fdf-a093-459f-9328-0549ef7b48c4" path="/var/lib/kubelet/pods/72d20fdf-a093-459f-9328-0549ef7b48c4/volumes" Mar 13 15:46:05 crc kubenswrapper[4907]: I0313 15:46:05.943601 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-7pld2"] Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.096644 4907 generic.go:334] "Generic (PLEG): container finished" podID="57610eac-b9eb-4628-bd99-b3ebb35b2772" containerID="27749e579eeb959a64f9197a7832237eb739e4daba11766edbfeb3f7aaf9ad37" exitCode=0 Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.096747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerDied","Data":"27749e579eeb959a64f9197a7832237eb739e4daba11766edbfeb3f7aaf9ad37"} Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.101199 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7pld2" event={"ID":"f6c0a640-5177-4188-abc3-d934f75624bb","Type":"ContainerStarted","Data":"f76a35ffcd772eb2bf027b892918eded8ae0d90b81519998bea60cc5f8343949"} Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.102842 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-p69c5" event={"ID":"ab5868a8-f083-4417-b7c0-f01dad538927","Type":"ContainerStarted","Data":"4400db9c9218563d608796b392e4cf7cc6279fc112b2adcdb10d47be37316c03"} Mar 13 15:46:06 crc kubenswrapper[4907]: I0313 15:46:06.858106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-hbbw4"] Mar 13 15:46:07 crc kubenswrapper[4907]: I0313 15:46:07.115265 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-hbbw4" event={"ID":"57610eac-b9eb-4628-bd99-b3ebb35b2772","Type":"ContainerStarted","Data":"1158feea7d437e31315b4d5ffcb87d298513a93189b92fc5e0a54854516e5572"} Mar 13 15:46:07 crc kubenswrapper[4907]: I0313 15:46:07.115533 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:07 crc kubenswrapper[4907]: I0313 15:46:07.145145 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-hbbw4" podStartSLOduration=5.145126576 podStartE2EDuration="5.145126576s" podCreationTimestamp="2026-03-13 15:46:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:46:07.136434938 +0000 UTC m=+6066.036222637" watchObservedRunningTime="2026-03-13 15:46:07.145126576 +0000 UTC m=+6066.044914265" Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.149822 4907 generic.go:334] "Generic (PLEG): container finished" podID="f6c0a640-5177-4188-abc3-d934f75624bb" containerID="f8136dbfa25da4d13b996f410c12861a563747986411d892b9b92b671cefafa6" exitCode=0 Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.149887 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7pld2" event={"ID":"f6c0a640-5177-4188-abc3-d934f75624bb","Type":"ContainerDied","Data":"f8136dbfa25da4d13b996f410c12861a563747986411d892b9b92b671cefafa6"} Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.152396 4907 generic.go:334] "Generic (PLEG): container finished" podID="ab5868a8-f083-4417-b7c0-f01dad538927" containerID="d83f400367245cd91cb0153fe29d0c6653531c070b4435e2eea12f7916668591" exitCode=0 Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.152640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-p69c5" event={"ID":"ab5868a8-f083-4417-b7c0-f01dad538927","Type":"ContainerDied","Data":"d83f400367245cd91cb0153fe29d0c6653531c070b4435e2eea12f7916668591"} Mar 13 15:46:09 crc kubenswrapper[4907]: I0313 15:46:09.782782 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:46:09 crc kubenswrapper[4907]: E0313 15:46:09.783569 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.163108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-7pld2" event={"ID":"f6c0a640-5177-4188-abc3-d934f75624bb","Type":"ContainerStarted","Data":"7d42d246473bf2b8e462bcf74b14d3e4997dd61f12af5cb13a8fb60e3bac48dd"} Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.164367 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.167026 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-p69c5" event={"ID":"ab5868a8-f083-4417-b7c0-f01dad538927","Type":"ContainerStarted","Data":"4bcbc2bf02d639dacf282548340a376215a25f0b29e4612a986c495aa5db94fd"} Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.167259 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.187307 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-7pld2" podStartSLOduration=4.42259988 podStartE2EDuration="6.187282843s" podCreationTimestamp="2026-03-13 15:46:04 +0000 UTC" firstStartedPulling="2026-03-13 15:46:05.968499869 +0000 UTC m=+6064.868287558" lastFinishedPulling="2026-03-13 15:46:07.733182832 +0000 UTC m=+6066.632970521" observedRunningTime="2026-03-13 15:46:10.18352672 +0000 UTC m=+6069.083314409" watchObservedRunningTime="2026-03-13 15:46:10.187282843 +0000 UTC m=+6069.087070532" Mar 13 15:46:10 crc kubenswrapper[4907]: I0313 15:46:10.210087 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-p69c5" podStartSLOduration=3.745419457 podStartE2EDuration="6.210068875s" podCreationTimestamp="2026-03-13 15:46:04 +0000 UTC" firstStartedPulling="2026-03-13 15:46:05.265751858 +0000 UTC m=+6064.165539547" lastFinishedPulling="2026-03-13 15:46:07.730401276 +0000 UTC m=+6066.630188965" observedRunningTime="2026-03-13 15:46:10.207867825 +0000 UTC m=+6069.107655524" watchObservedRunningTime="2026-03-13 15:46:10.210068875 +0000 UTC m=+6069.109856554" Mar 13 15:46:17 crc kubenswrapper[4907]: I0313 15:46:17.913726 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-hbbw4" Mar 13 15:46:19 crc kubenswrapper[4907]: I0313 15:46:19.713081 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-p69c5" Mar 13 15:46:20 crc kubenswrapper[4907]: I0313 15:46:20.384313 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-7pld2" Mar 13 15:46:21 crc kubenswrapper[4907]: I0313 15:46:21.790206 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:46:22 crc kubenswrapper[4907]: I0313 15:46:22.278469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2"} Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.272852 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.275814 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.280996 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.281027 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.281318 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.281394 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-tpjbz" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.289630 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336006 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336202 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.336254 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.342058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.342313 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" containerID="cri-o://56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.342732 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" containerID="cri-o://d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.411998 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.412549 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" containerID="cri-o://02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.412711 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" containerID="cri-o://27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f" gracePeriod=30 Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.437977 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438052 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438178 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.438666 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.439751 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.439801 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.450727 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.470493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"horizon-c7977bcd9-2vs8h\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.480257 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.481982 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.520159 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540357 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.540518 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.605324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642714 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642774 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642817 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.642927 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.643717 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.644392 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.644481 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.647183 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.670729 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"horizon-769b54bfcc-kbtds\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:26 crc kubenswrapper[4907]: I0313 15:46:26.906198 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.078179 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.123293 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.124912 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.146239 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.151850 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.152192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.152579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.152931 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.153037 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: W0313 15:46:27.159258 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9fbab8_539c_48f6_8cae_3eda23300b28.slice/crio-941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d WatchSource:0}: Error finding container 941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d: Status 404 returned error can't find the container with id 941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.180000 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255462 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255521 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.255590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.257141 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.258285 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.258459 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.262408 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.276235 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"horizon-655cbf768f-dm9zz\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.336044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerStarted","Data":"941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d"} Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.338730 4907 generic.go:334] "Generic (PLEG): container finished" podID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerID="56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85" exitCode=143 Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.338836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerDied","Data":"56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85"} Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.341556 4907 generic.go:334] "Generic (PLEG): container finished" podID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerID="02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8" exitCode=143 Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.341585 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerDied","Data":"02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8"} Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.486452 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.487729 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:46:27 crc kubenswrapper[4907]: W0313 15:46:27.489103 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb WatchSource:0}: Error finding container e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb: Status 404 returned error can't find the container with id e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb Mar 13 15:46:27 crc kubenswrapper[4907]: W0313 15:46:27.946233 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafd4001d_40a5_4a09_ab05_36cbd4d06eb4.slice/crio-8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365 WatchSource:0}: Error finding container 8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365: Status 404 returned error can't find the container with id 8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365 Mar 13 15:46:27 crc kubenswrapper[4907]: I0313 15:46:27.948254 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:46:28 crc kubenswrapper[4907]: I0313 15:46:28.355980 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerStarted","Data":"e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb"} Mar 13 15:46:28 crc kubenswrapper[4907]: I0313 15:46:28.358001 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerStarted","Data":"8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365"} Mar 13 15:46:29 crc kubenswrapper[4907]: I0313 15:46:29.740267 4907 scope.go:117] "RemoveContainer" containerID="3f861dd2c3ff8fa5314ce9c35d45445320ad82d6381b4138837ad16c289319f5" Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.382417 4907 generic.go:334] "Generic (PLEG): container finished" podID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerID="27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f" exitCode=0 Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.382498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerDied","Data":"27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f"} Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.387216 4907 generic.go:334] "Generic (PLEG): container finished" podID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerID="d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03" exitCode=0 Mar 13 15:46:30 crc kubenswrapper[4907]: I0313 15:46:30.387253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerDied","Data":"d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03"} Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.649367 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.655225 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697198 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697288 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697314 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697347 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697429 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697487 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697566 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697600 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697682 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697725 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697769 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697817 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") pod \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\" (UID: \"096d3713-ea4f-43d3-bfb9-9170e8958ed0\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697870 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.697936 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") pod \"86794caf-bf6d-4394-bc1c-3c516bdbada8\" (UID: \"86794caf-bf6d-4394-bc1c-3c516bdbada8\") " Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.708287 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs" (OuterVolumeSpecName: "logs") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.708823 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.711262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts" (OuterVolumeSpecName: "scripts") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.711787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts" (OuterVolumeSpecName: "scripts") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.711859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs" (OuterVolumeSpecName: "logs") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.712152 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph" (OuterVolumeSpecName: "ceph") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.718723 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn" (OuterVolumeSpecName: "kube-api-access-4srpn") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "kube-api-access-4srpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.719302 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph" (OuterVolumeSpecName: "ceph") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.723431 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.729601 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5" (OuterVolumeSpecName: "kube-api-access-h7zl5") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "kube-api-access-h7zl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.737316 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.750450 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.776996 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data" (OuterVolumeSpecName: "config-data") pod "86794caf-bf6d-4394-bc1c-3c516bdbada8" (UID: "86794caf-bf6d-4394-bc1c-3c516bdbada8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.785630 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data" (OuterVolumeSpecName: "config-data") pod "096d3713-ea4f-43d3-bfb9-9170e8958ed0" (UID: "096d3713-ea4f-43d3-bfb9-9170e8958ed0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799729 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7zl5\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-kube-api-access-h7zl5\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799756 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799765 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799773 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799783 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799791 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/096d3713-ea4f-43d3-bfb9-9170e8958ed0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799801 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86794caf-bf6d-4394-bc1c-3c516bdbada8-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799809 4907 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799817 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799825 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86794caf-bf6d-4394-bc1c-3c516bdbada8-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799833 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/096d3713-ea4f-43d3-bfb9-9170e8958ed0-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799842 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4srpn\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-kube-api-access-4srpn\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799850 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/86794caf-bf6d-4394-bc1c-3c516bdbada8-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:34 crc kubenswrapper[4907]: I0313 15:46:34.799858 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/096d3713-ea4f-43d3-bfb9-9170e8958ed0-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.434964 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerStarted","Data":"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.435520 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerStarted","Data":"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.438229 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"86794caf-bf6d-4394-bc1c-3c516bdbada8","Type":"ContainerDied","Data":"e6d5e941f6575995d025c2bf565c8da48222d14bbb2fbb1710f7cd79e1394fae"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.438254 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.438328 4907 scope.go:117] "RemoveContainer" containerID="d5e6285fe8e2b22da125d5022a7924d7086ba68515ab90ca41e35d0607040a03" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.441961 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"096d3713-ea4f-43d3-bfb9-9170e8958ed0","Type":"ContainerDied","Data":"13bb6e3535193b961d70131bf3aebbe0d658a8a82ddb3d84795b101e8ae9b66a"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.442139 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.451546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerStarted","Data":"862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.451598 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerStarted","Data":"fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.451758 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-769b54bfcc-kbtds" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" containerID="cri-o://fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd" gracePeriod=30 Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.452105 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-769b54bfcc-kbtds" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" containerID="cri-o://862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151" gracePeriod=30 Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.457832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerStarted","Data":"b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.457865 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerStarted","Data":"4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f"} Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.473405 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c7977bcd9-2vs8h" podStartSLOduration=1.85554039 podStartE2EDuration="9.473387163s" podCreationTimestamp="2026-03-13 15:46:26 +0000 UTC" firstStartedPulling="2026-03-13 15:46:27.1744045 +0000 UTC m=+6086.074192189" lastFinishedPulling="2026-03-13 15:46:34.792251273 +0000 UTC m=+6093.692038962" observedRunningTime="2026-03-13 15:46:35.452509902 +0000 UTC m=+6094.352297591" watchObservedRunningTime="2026-03-13 15:46:35.473387163 +0000 UTC m=+6094.373174852" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.480274 4907 scope.go:117] "RemoveContainer" containerID="56ff9c5433a9cbee0386dea3d88b4e88f870fa5dcf67d5221e12245f174e8f85" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.491085 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-769b54bfcc-kbtds" podStartSLOduration=2.149175853 podStartE2EDuration="9.491046555s" podCreationTimestamp="2026-03-13 15:46:26 +0000 UTC" firstStartedPulling="2026-03-13 15:46:27.504369156 +0000 UTC m=+6086.404156845" lastFinishedPulling="2026-03-13 15:46:34.846239858 +0000 UTC m=+6093.746027547" observedRunningTime="2026-03-13 15:46:35.477732611 +0000 UTC m=+6094.377520340" watchObservedRunningTime="2026-03-13 15:46:35.491046555 +0000 UTC m=+6094.390834244" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.504671 4907 scope.go:117] "RemoveContainer" containerID="27209680abb804b6c48ccdaa33b44ab4f54951d1df1401058af333f852827d0f" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.504821 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.514522 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.533009 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.533967 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534113 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.534227 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534309 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.534418 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534517 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: E0313 15:46:35.534620 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.534695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535072 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535419 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-log" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.535499 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" containerName="glance-httpd" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.536002 4907 scope.go:117] "RemoveContainer" containerID="02ef5ab885e1e984508608add8cbc5f616e0cf1ed8d42cf7ded2960a6fa681f8" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.537249 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.537284 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-655cbf768f-dm9zz" podStartSLOduration=1.738425962 podStartE2EDuration="8.537263538s" podCreationTimestamp="2026-03-13 15:46:27 +0000 UTC" firstStartedPulling="2026-03-13 15:46:27.951089811 +0000 UTC m=+6086.850877500" lastFinishedPulling="2026-03-13 15:46:34.749927397 +0000 UTC m=+6093.649715076" observedRunningTime="2026-03-13 15:46:35.522764982 +0000 UTC m=+6094.422552671" watchObservedRunningTime="2026-03-13 15:46:35.537263538 +0000 UTC m=+6094.437051227" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.551839 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.552089 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.552547 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5c589" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.622010 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnn54\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-kube-api-access-qnn54\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623001 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-logs\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623124 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623431 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623562 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623675 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.623770 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.653382 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.664559 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.678596 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.695382 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.697495 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.703001 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.726011 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnn54\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-kube-api-access-qnn54\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727547 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727624 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727659 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-logs\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727691 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727712 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-logs\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.727782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728621 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728659 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728688 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728711 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvfdw\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-kube-api-access-mvfdw\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.728753 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-logs\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.729290 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/70b550c7-5c28-416d-83c3-1ba63ecd5abd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.734502 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-ceph\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.735795 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.745931 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.747610 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70b550c7-5c28-416d-83c3-1ba63ecd5abd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.751014 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnn54\" (UniqueName: \"kubernetes.io/projected/70b550c7-5c28-416d-83c3-1ba63ecd5abd-kube-api-access-qnn54\") pod \"glance-default-internal-api-0\" (UID: \"70b550c7-5c28-416d-83c3-1ba63ecd5abd\") " pod="openstack/glance-default-internal-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.794205 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="096d3713-ea4f-43d3-bfb9-9170e8958ed0" path="/var/lib/kubelet/pods/096d3713-ea4f-43d3-bfb9-9170e8958ed0/volumes" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.795318 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86794caf-bf6d-4394-bc1c-3c516bdbada8" path="/var/lib/kubelet/pods/86794caf-bf6d-4394-bc1c-3c516bdbada8/volumes" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830577 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830645 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830750 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830793 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvfdw\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-kube-api-access-mvfdw\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-logs\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.830871 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.831373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.835630 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.836528 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95435ffc-5322-4bd8-9558-421f953096eb-logs\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.837161 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.837373 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.838314 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95435ffc-5322-4bd8-9558-421f953096eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.855948 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvfdw\" (UniqueName: \"kubernetes.io/projected/95435ffc-5322-4bd8-9558-421f953096eb-kube-api-access-mvfdw\") pod \"glance-default-external-api-0\" (UID: \"95435ffc-5322-4bd8-9558-421f953096eb\") " pod="openstack/glance-default-external-api-0" Mar 13 15:46:35 crc kubenswrapper[4907]: I0313 15:46:35.949294 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.014423 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.531013 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 15:46:36 crc kubenswrapper[4907]: W0313 15:46:36.532107 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70b550c7_5c28_416d_83c3_1ba63ecd5abd.slice/crio-1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c WatchSource:0}: Error finding container 1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c: Status 404 returned error can't find the container with id 1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.606212 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.606266 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.677532 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 15:46:36 crc kubenswrapper[4907]: I0313 15:46:36.907392 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.487015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.487333 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.506700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"70b550c7-5c28-416d-83c3-1ba63ecd5abd","Type":"ContainerStarted","Data":"36f29b0ba8237aac66775608213473b8361269473dc952d84e52449563485b76"} Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.506755 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"70b550c7-5c28-416d-83c3-1ba63ecd5abd","Type":"ContainerStarted","Data":"1efed8d64e814a0df83522691cc4cb2688b670fd49736ac116a8018a6049120c"} Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.509645 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95435ffc-5322-4bd8-9558-421f953096eb","Type":"ContainerStarted","Data":"9db337a620f677125ecfbd67707c679d5ec92247b100824435ccaca12a038a56"} Mar 13 15:46:37 crc kubenswrapper[4907]: I0313 15:46:37.509801 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95435ffc-5322-4bd8-9558-421f953096eb","Type":"ContainerStarted","Data":"9ad8613388673a254440d120f808de8b4cb2ead90a90acf7bbca778db8b02af1"} Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.519736 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"95435ffc-5322-4bd8-9558-421f953096eb","Type":"ContainerStarted","Data":"8a4d44e895e491e861e19682edef50a39dd9abc89004a2a6cb88ed9b4b4b3e21"} Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.522305 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"70b550c7-5c28-416d-83c3-1ba63ecd5abd","Type":"ContainerStarted","Data":"80d182356297060953e417203d7ce12f38371db0d77c321425cf2bb8d290165d"} Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.577806 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.577783161 podStartE2EDuration="3.577783161s" podCreationTimestamp="2026-03-13 15:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:46:38.567684934 +0000 UTC m=+6097.467472623" watchObservedRunningTime="2026-03-13 15:46:38.577783161 +0000 UTC m=+6097.477570850" Mar 13 15:46:38 crc kubenswrapper[4907]: I0313 15:46:38.578269 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.578263393 podStartE2EDuration="3.578263393s" podCreationTimestamp="2026-03-13 15:46:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:46:38.550523745 +0000 UTC m=+6097.450311434" watchObservedRunningTime="2026-03-13 15:46:38.578263393 +0000 UTC m=+6097.478051082" Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.041223 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.050003 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.058911 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4jnct"] Mar 13 15:46:44 crc kubenswrapper[4907]: I0313 15:46:44.066714 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b654-account-create-update-4jjkj"] Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.798603 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e73852b-fdd0-4cd2-a268-96f6ed551011" path="/var/lib/kubelet/pods/1e73852b-fdd0-4cd2-a268-96f6ed551011/volumes" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.800224 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="523d3b7c-7098-44bc-92a6-d48f3ae49b65" path="/var/lib/kubelet/pods/523d3b7c-7098-44bc-92a6-d48f3ae49b65/volumes" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.950483 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.950551 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.987410 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:45 crc kubenswrapper[4907]: I0313 15:46:45.990954 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.015285 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.015353 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.054287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.066742 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600332 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600381 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600397 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.600408 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:46 crc kubenswrapper[4907]: I0313 15:46:46.617829 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:46:47 crc kubenswrapper[4907]: I0313 15:46:47.490363 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.930154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.930494 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.934294 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.934386 4907 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.934809 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 15:46:48 crc kubenswrapper[4907]: I0313 15:46:48.953861 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 15:46:50 crc kubenswrapper[4907]: I0313 15:46:50.034092 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:46:50 crc kubenswrapper[4907]: I0313 15:46:50.044198 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xrjlt"] Mar 13 15:46:51 crc kubenswrapper[4907]: I0313 15:46:51.824907 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df0d68f5-c073-4dff-8f70-f2897aaef9dc" path="/var/lib/kubelet/pods/df0d68f5-c073-4dff-8f70-f2897aaef9dc/volumes" Mar 13 15:46:58 crc kubenswrapper[4907]: I0313 15:46:58.418937 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:46:59 crc kubenswrapper[4907]: I0313 15:46:59.335510 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:47:00 crc kubenswrapper[4907]: I0313 15:47:00.030061 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.140310 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.217549 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.218101 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" containerID="cri-o://d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" gracePeriod=30 Mar 13 15:47:01 crc kubenswrapper[4907]: I0313 15:47:01.218200 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" containerID="cri-o://64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" gracePeriod=30 Mar 13 15:47:04 crc kubenswrapper[4907]: I0313 15:47:04.781706 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" exitCode=0 Mar 13 15:47:04 crc kubenswrapper[4907]: I0313 15:47:04.781768 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerDied","Data":"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c"} Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795338 4907 generic.go:334] "Generic (PLEG): container finished" podID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerID="862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151" exitCode=137 Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795629 4907 generic.go:334] "Generic (PLEG): container finished" podID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerID="fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd" exitCode=137 Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795625 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerDied","Data":"862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151"} Mar 13 15:47:05 crc kubenswrapper[4907]: I0313 15:47:05.795670 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerDied","Data":"fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd"} Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.062378 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159221 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159291 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159562 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159634 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159817 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") pod \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\" (UID: \"099b49c3-666e-4bf5-9c15-32cf9c40e1e7\") " Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.159981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs" (OuterVolumeSpecName: "logs") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.160531 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.166252 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb" (OuterVolumeSpecName: "kube-api-access-scsbb") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "kube-api-access-scsbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.166263 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.203939 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data" (OuterVolumeSpecName: "config-data") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.205402 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts" (OuterVolumeSpecName: "scripts") pod "099b49c3-666e-4bf5-9c15-32cf9c40e1e7" (UID: "099b49c3-666e-4bf5-9c15-32cf9c40e1e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.262950 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.262995 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.263007 4907 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.263020 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scsbb\" (UniqueName: \"kubernetes.io/projected/099b49c3-666e-4bf5-9c15-32cf9c40e1e7-kube-api-access-scsbb\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.606168 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.804794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-769b54bfcc-kbtds" event={"ID":"099b49c3-666e-4bf5-9c15-32cf9c40e1e7","Type":"ContainerDied","Data":"e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb"} Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.804839 4907 scope.go:117] "RemoveContainer" containerID="862291f2b2864cad1d39537f26d6a5251dddefb5e909a977e483fcdb34432151" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.804963 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-769b54bfcc-kbtds" Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.847674 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.855543 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-769b54bfcc-kbtds"] Mar 13 15:47:06 crc kubenswrapper[4907]: I0313 15:47:06.984711 4907 scope.go:117] "RemoveContainer" containerID="fcce68c0980dbae069e759717a0e2db31c52e1f7a61e3feb3e5debebb904b5dd" Mar 13 15:47:07 crc kubenswrapper[4907]: I0313 15:47:07.807311 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" path="/var/lib/kubelet/pods/099b49c3-666e-4bf5-9c15-32cf9c40e1e7/volumes" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.546652 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d79dff969-ssj4t"] Mar 13 15:47:08 crc kubenswrapper[4907]: E0313 15:47:08.547425 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547447 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" Mar 13 15:47:08 crc kubenswrapper[4907]: E0313 15:47:08.547492 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547502 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547737 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon-log" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.547774 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="099b49c3-666e-4bf5-9c15-32cf9c40e1e7" containerName="horizon" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.549238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.562400 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d79dff969-ssj4t"] Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tsd8\" (UniqueName: \"kubernetes.io/projected/5315ef54-fd95-400c-a7ef-a356c0f867bc-kube-api-access-8tsd8\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615579 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-scripts\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-config-data\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615629 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5315ef54-fd95-400c-a7ef-a356c0f867bc-horizon-secret-key\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.615755 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5315ef54-fd95-400c-a7ef-a356c0f867bc-logs\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717035 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5315ef54-fd95-400c-a7ef-a356c0f867bc-logs\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717146 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tsd8\" (UniqueName: \"kubernetes.io/projected/5315ef54-fd95-400c-a7ef-a356c0f867bc-kube-api-access-8tsd8\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717188 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-scripts\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717206 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-config-data\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5315ef54-fd95-400c-a7ef-a356c0f867bc-horizon-secret-key\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.717737 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5315ef54-fd95-400c-a7ef-a356c0f867bc-logs\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.718171 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-scripts\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.718983 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5315ef54-fd95-400c-a7ef-a356c0f867bc-config-data\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.722069 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5315ef54-fd95-400c-a7ef-a356c0f867bc-horizon-secret-key\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.741262 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tsd8\" (UniqueName: \"kubernetes.io/projected/5315ef54-fd95-400c-a7ef-a356c0f867bc-kube-api-access-8tsd8\") pod \"horizon-6d79dff969-ssj4t\" (UID: \"5315ef54-fd95-400c-a7ef-a356c0f867bc\") " pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:08 crc kubenswrapper[4907]: I0313 15:47:08.870133 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.344547 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d79dff969-ssj4t"] Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.838126 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d79dff969-ssj4t" event={"ID":"5315ef54-fd95-400c-a7ef-a356c0f867bc","Type":"ContainerStarted","Data":"d662824b3d4144a85a7254045e81fec495afd2d985d3e17becf10c307f110f3f"} Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.838478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d79dff969-ssj4t" event={"ID":"5315ef54-fd95-400c-a7ef-a356c0f867bc","Type":"ContainerStarted","Data":"869c9c9925a981e0b81d87611070cdcfe490596c4315540b6e041d946c3a07dc"} Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.838499 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d79dff969-ssj4t" event={"ID":"5315ef54-fd95-400c-a7ef-a356c0f867bc","Type":"ContainerStarted","Data":"9c2349bc5006f4b36b5615d4e495b3a166fcd8e0f8d0962d4816a5e2a2c9a9a3"} Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.859633 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6d79dff969-ssj4t" podStartSLOduration=1.859611913 podStartE2EDuration="1.859611913s" podCreationTimestamp="2026-03-13 15:47:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:47:09.857254779 +0000 UTC m=+6128.757042488" watchObservedRunningTime="2026-03-13 15:47:09.859611913 +0000 UTC m=+6128.759399602" Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.995160 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:47:09 crc kubenswrapper[4907]: I0313 15:47:09.997429 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.017613 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.116643 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.131001 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.131333 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.134590 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.160584 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.161054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268871 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268961 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.268987 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.269743 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.298860 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"heat-db-create-lnsws\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.357764 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.371100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.371412 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.372122 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.387250 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"heat-019b-account-create-update-848sw\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:10 crc kubenswrapper[4907]: I0313 15:47:10.451998 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.041025 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.137705 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.859320 4907 generic.go:334] "Generic (PLEG): container finished" podID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerID="839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880" exitCode=0 Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.859375 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-019b-account-create-update-848sw" event={"ID":"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49","Type":"ContainerDied","Data":"839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880"} Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.859424 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-019b-account-create-update-848sw" event={"ID":"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49","Type":"ContainerStarted","Data":"d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71"} Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.861264 4907 generic.go:334] "Generic (PLEG): container finished" podID="490bd531-09f6-48e5-b194-b79e9b532131" containerID="daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486" exitCode=0 Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.861314 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lnsws" event={"ID":"490bd531-09f6-48e5-b194-b79e9b532131","Type":"ContainerDied","Data":"daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486"} Mar 13 15:47:11 crc kubenswrapper[4907]: I0313 15:47:11.861337 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lnsws" event={"ID":"490bd531-09f6-48e5-b194-b79e9b532131","Type":"ContainerStarted","Data":"23b375d293ca9398feacd198bd47f6604e6d9e7963768ae62f0d245f7da436e3"} Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.344617 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.360989 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.369499 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") pod \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.370723 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") pod \"490bd531-09f6-48e5-b194-b79e9b532131\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.370940 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") pod \"490bd531-09f6-48e5-b194-b79e9b532131\" (UID: \"490bd531-09f6-48e5-b194-b79e9b532131\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.371102 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") pod \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\" (UID: \"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49\") " Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.371774 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "490bd531-09f6-48e5-b194-b79e9b532131" (UID: "490bd531-09f6-48e5-b194-b79e9b532131"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.375320 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" (UID: "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.378783 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms" (OuterVolumeSpecName: "kube-api-access-khvms") pod "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" (UID: "ce0b08b9-c4e1-4236-8f54-cf78b45d2d49"). InnerVolumeSpecName "kube-api-access-khvms". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.388589 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn" (OuterVolumeSpecName: "kube-api-access-wx7rn") pod "490bd531-09f6-48e5-b194-b79e9b532131" (UID: "490bd531-09f6-48e5-b194-b79e9b532131"). InnerVolumeSpecName "kube-api-access-wx7rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.389774 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.389911 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khvms\" (UniqueName: \"kubernetes.io/projected/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49-kube-api-access-khvms\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.389925 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/490bd531-09f6-48e5-b194-b79e9b532131-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.390034 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wx7rn\" (UniqueName: \"kubernetes.io/projected/490bd531-09f6-48e5-b194-b79e9b532131-kube-api-access-wx7rn\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.886553 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-019b-account-create-update-848sw" event={"ID":"ce0b08b9-c4e1-4236-8f54-cf78b45d2d49","Type":"ContainerDied","Data":"d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71"} Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.886601 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.886671 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-019b-account-create-update-848sw" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.889486 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-lnsws" event={"ID":"490bd531-09f6-48e5-b194-b79e9b532131","Type":"ContainerDied","Data":"23b375d293ca9398feacd198bd47f6604e6d9e7963768ae62f0d245f7da436e3"} Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.889516 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b375d293ca9398feacd198bd47f6604e6d9e7963768ae62f0d245f7da436e3" Mar 13 15:47:13 crc kubenswrapper[4907]: I0313 15:47:13.889657 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-lnsws" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.326212 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:47:15 crc kubenswrapper[4907]: E0313 15:47:15.327135 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerName="mariadb-account-create-update" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327153 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerName="mariadb-account-create-update" Mar 13 15:47:15 crc kubenswrapper[4907]: E0313 15:47:15.327221 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="490bd531-09f6-48e5-b194-b79e9b532131" containerName="mariadb-database-create" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327231 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="490bd531-09f6-48e5-b194-b79e9b532131" containerName="mariadb-database-create" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327484 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" containerName="mariadb-account-create-update" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.327505 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="490bd531-09f6-48e5-b194-b79e9b532131" containerName="mariadb-database-create" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.328439 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.332160 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.332653 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-mcdkh" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.339027 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.429016 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.429080 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.429449 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.531838 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.531903 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.531962 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.545922 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.546079 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.548287 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"heat-db-sync-mqnk8\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:15 crc kubenswrapper[4907]: I0313 15:47:15.649331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:16 crc kubenswrapper[4907]: I0313 15:47:16.165412 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:47:16 crc kubenswrapper[4907]: I0313 15:47:16.606996 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:47:16 crc kubenswrapper[4907]: I0313 15:47:16.920864 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerStarted","Data":"3d55337bfec3fbfeaf31255470fe615e54c37d18887cd6469654963c1c88fe12"} Mar 13 15:47:17 crc kubenswrapper[4907]: I0313 15:47:17.975420 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:17 crc kubenswrapper[4907]: I0313 15:47:17.979685 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.012166 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.055862 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.066875 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.076423 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-jh7h4"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.085286 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-629b-account-create-update-rptv2"] Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.086815 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.086965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.087324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.189426 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.189589 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.189656 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.190578 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.190904 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.222096 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"community-operators-tthk2\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.317724 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.871205 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.871484 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:18 crc kubenswrapper[4907]: I0313 15:47:18.942160 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.799309 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cbbccb7-b0c0-49be-a969-ffa18b07bd13" path="/var/lib/kubelet/pods/3cbbccb7-b0c0-49be-a969-ffa18b07bd13/volumes" Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.800726 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="608c9419-22d5-4915-b5c6-8be70eebdb23" path="/var/lib/kubelet/pods/608c9419-22d5-4915-b5c6-8be70eebdb23/volumes" Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.963669 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" exitCode=0 Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.963762 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916"} Mar 13 15:47:19 crc kubenswrapper[4907]: I0313 15:47:19.963798 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerStarted","Data":"ab17e9b088419fabf1ec951208897a9f96839403f2d2331918d09059ad41ce73"} Mar 13 15:47:25 crc kubenswrapper[4907]: I0313 15:47:25.017586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerStarted","Data":"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff"} Mar 13 15:47:25 crc kubenswrapper[4907]: I0313 15:47:25.021017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerStarted","Data":"28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833"} Mar 13 15:47:25 crc kubenswrapper[4907]: I0313 15:47:25.072727 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-mqnk8" podStartSLOduration=1.6402556719999999 podStartE2EDuration="10.072706601s" podCreationTimestamp="2026-03-13 15:47:15 +0000 UTC" firstStartedPulling="2026-03-13 15:47:16.165783898 +0000 UTC m=+6135.065571587" lastFinishedPulling="2026-03-13 15:47:24.598234827 +0000 UTC m=+6143.498022516" observedRunningTime="2026-03-13 15:47:25.063248522 +0000 UTC m=+6143.963036211" watchObservedRunningTime="2026-03-13 15:47:25.072706601 +0000 UTC m=+6143.972494280" Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.033227 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" exitCode=0 Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.033271 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff"} Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.606825 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c7977bcd9-2vs8h" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.154:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.154:8080: connect: connection refused" Mar 13 15:47:26 crc kubenswrapper[4907]: I0313 15:47:26.607049 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.037451 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.048020 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerStarted","Data":"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348"} Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.048726 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jtsxr"] Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.072539 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tthk2" podStartSLOduration=6.958694294 podStartE2EDuration="10.072519259s" podCreationTimestamp="2026-03-13 15:47:17 +0000 UTC" firstStartedPulling="2026-03-13 15:47:23.666287335 +0000 UTC m=+6142.566075024" lastFinishedPulling="2026-03-13 15:47:26.7801123 +0000 UTC m=+6145.679899989" observedRunningTime="2026-03-13 15:47:27.065356883 +0000 UTC m=+6145.965144572" watchObservedRunningTime="2026-03-13 15:47:27.072519259 +0000 UTC m=+6145.972306948" Mar 13 15:47:27 crc kubenswrapper[4907]: I0313 15:47:27.793171 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af3ec26-0a7c-44c4-8073-fc93093c0cf7" path="/var/lib/kubelet/pods/1af3ec26-0a7c-44c4-8073-fc93093c0cf7/volumes" Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.058148 4907 generic.go:334] "Generic (PLEG): container finished" podID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerID="28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833" exitCode=0 Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.058232 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerDied","Data":"28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833"} Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.318276 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.318323 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:28 crc kubenswrapper[4907]: I0313 15:47:28.873665 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6d79dff969-ssj4t" podUID="5315ef54-fd95-400c-a7ef-a356c0f867bc" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.159:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.159:8080: connect: connection refused" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.394435 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tthk2" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" probeResult="failure" output=< Mar 13 15:47:29 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:47:29 crc kubenswrapper[4907]: > Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.425319 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.556466 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") pod \"6df151b5-1c14-488c-9f23-9d6d629b493d\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.556625 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") pod \"6df151b5-1c14-488c-9f23-9d6d629b493d\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.556670 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") pod \"6df151b5-1c14-488c-9f23-9d6d629b493d\" (UID: \"6df151b5-1c14-488c-9f23-9d6d629b493d\") " Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.562335 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm" (OuterVolumeSpecName: "kube-api-access-jzrmm") pod "6df151b5-1c14-488c-9f23-9d6d629b493d" (UID: "6df151b5-1c14-488c-9f23-9d6d629b493d"). InnerVolumeSpecName "kube-api-access-jzrmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.595904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6df151b5-1c14-488c-9f23-9d6d629b493d" (UID: "6df151b5-1c14-488c-9f23-9d6d629b493d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.639600 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data" (OuterVolumeSpecName: "config-data") pod "6df151b5-1c14-488c-9f23-9d6d629b493d" (UID: "6df151b5-1c14-488c-9f23-9d6d629b493d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.659153 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.659191 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df151b5-1c14-488c-9f23-9d6d629b493d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:29 crc kubenswrapper[4907]: I0313 15:47:29.659213 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzrmm\" (UniqueName: \"kubernetes.io/projected/6df151b5-1c14-488c-9f23-9d6d629b493d-kube-api-access-jzrmm\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:30 crc kubenswrapper[4907]: I0313 15:47:30.077300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-mqnk8" event={"ID":"6df151b5-1c14-488c-9f23-9d6d629b493d","Type":"ContainerDied","Data":"3d55337bfec3fbfeaf31255470fe615e54c37d18887cd6469654963c1c88fe12"} Mar 13 15:47:30 crc kubenswrapper[4907]: I0313 15:47:30.077634 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d55337bfec3fbfeaf31255470fe615e54c37d18887cd6469654963c1c88fe12" Mar 13 15:47:30 crc kubenswrapper[4907]: I0313 15:47:30.077376 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-mqnk8" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.269935 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-86cbbf5dd8-25c77"] Mar 13 15:47:31 crc kubenswrapper[4907]: E0313 15:47:31.270383 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerName="heat-db-sync" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.270394 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerName="heat-db-sync" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.270596 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" containerName="heat-db-sync" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.271283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.282015 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.282323 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.282488 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-mcdkh" Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293569 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-d2ee4f22d00d1f9ceaf1d6575365b027f64e73a1473c8ba545cf5581cfc37e71: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293642 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-conmon-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-conmon-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293665 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod490bd531_09f6_48e5_b194_b79e9b532131.slice/crio-daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod490bd531_09f6_48e5_b194_b79e9b532131.slice/crio-daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486.scope: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293680 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce0b08b9_c4e1_4236_8f54_cf78b45d2d49.slice/crio-839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880.scope: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: W0313 15:47:31.293754 4907 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df151b5_1c14_488c_9f23_9d6d629b493d.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df151b5_1c14_488c_9f23_9d6d629b493d.slice: no such file or directory Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.321054 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-86cbbf5dd8-25c77"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.383856 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7fd69496d-xglm5"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390062 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data-custom\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390163 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-combined-ca-bundle\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390226 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfmqz\" (UniqueName: \"kubernetes.io/projected/3dd72cd1-7342-40aa-8c3c-9d14375b3930-kube-api-access-jfmqz\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.390430 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.405243 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fd69496d-xglm5"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.405366 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.412264 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.426996 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7dbbf4d5b8-846ft"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.428230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.429935 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.443549 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7dbbf4d5b8-846ft"] Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493286 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493380 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493418 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-combined-ca-bundle\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493435 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-combined-ca-bundle\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493466 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data-custom\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493487 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data-custom\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493524 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-combined-ca-bundle\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8c5v\" (UniqueName: \"kubernetes.io/projected/1d7cc945-4d68-41c8-9ea0-359af3279cc8-kube-api-access-p8c5v\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493593 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw9pz\" (UniqueName: \"kubernetes.io/projected/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-kube-api-access-cw9pz\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493616 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfmqz\" (UniqueName: \"kubernetes.io/projected/3dd72cd1-7342-40aa-8c3c-9d14375b3930-kube-api-access-jfmqz\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.493638 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data-custom\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.514684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.515599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-config-data-custom\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.519680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfmqz\" (UniqueName: \"kubernetes.io/projected/3dd72cd1-7342-40aa-8c3c-9d14375b3930-kube-api-access-jfmqz\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.526661 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dd72cd1-7342-40aa-8c3c-9d14375b3930-combined-ca-bundle\") pod \"heat-engine-86cbbf5dd8-25c77\" (UID: \"3dd72cd1-7342-40aa-8c3c-9d14375b3930\") " pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.596856 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597059 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-combined-ca-bundle\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597085 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-combined-ca-bundle\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597155 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data-custom\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597272 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8c5v\" (UniqueName: \"kubernetes.io/projected/1d7cc945-4d68-41c8-9ea0-359af3279cc8-kube-api-access-p8c5v\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597319 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw9pz\" (UniqueName: \"kubernetes.io/projected/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-kube-api-access-cw9pz\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.597355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data-custom\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.604786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data-custom\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.604853 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.608976 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-config-data-custom\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.614707 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-config-data\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.616695 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8c5v\" (UniqueName: \"kubernetes.io/projected/1d7cc945-4d68-41c8-9ea0-359af3279cc8-kube-api-access-p8c5v\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.623843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-combined-ca-bundle\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.623928 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d7cc945-4d68-41c8-9ea0-359af3279cc8-combined-ca-bundle\") pod \"heat-cfnapi-7dbbf4d5b8-846ft\" (UID: \"1d7cc945-4d68-41c8-9ea0-359af3279cc8\") " pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.630723 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw9pz\" (UniqueName: \"kubernetes.io/projected/a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a-kube-api-access-cw9pz\") pod \"heat-api-7fd69496d-xglm5\" (UID: \"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a\") " pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: E0313 15:47:31.641846 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9fbab8_539c_48f6_8cae_3eda23300b28.slice/crio-conmon-d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d9fbab8_539c_48f6_8cae_3eda23300b28.slice/crio-d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.726016 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.754982 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.767922 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.776486 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904002 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904094 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904201 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904297 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904418 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") pod \"2d9fbab8-539c-48f6-8cae-3eda23300b28\" (UID: \"2d9fbab8-539c-48f6-8cae-3eda23300b28\") " Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.904607 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs" (OuterVolumeSpecName: "logs") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.905015 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d9fbab8-539c-48f6-8cae-3eda23300b28-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.910009 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.918778 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq" (OuterVolumeSpecName: "kube-api-access-bjhcq") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "kube-api-access-bjhcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.944909 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts" (OuterVolumeSpecName: "scripts") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:31 crc kubenswrapper[4907]: I0313 15:47:31.979093 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data" (OuterVolumeSpecName: "config-data") pod "2d9fbab8-539c-48f6-8cae-3eda23300b28" (UID: "2d9fbab8-539c-48f6-8cae-3eda23300b28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007400 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007443 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjhcq\" (UniqueName: \"kubernetes.io/projected/2d9fbab8-539c-48f6-8cae-3eda23300b28-kube-api-access-bjhcq\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007543 4907 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2d9fbab8-539c-48f6-8cae-3eda23300b28-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.007556 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d9fbab8-539c-48f6-8cae-3eda23300b28-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099143 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" exitCode=137 Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099187 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerDied","Data":"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29"} Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099211 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c7977bcd9-2vs8h" event={"ID":"2d9fbab8-539c-48f6-8cae-3eda23300b28","Type":"ContainerDied","Data":"941b229bb7806138fc500a8ef3b0ed37c2227cce03ef0b5260159ccb8082332d"} Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099226 4907 scope.go:117] "RemoveContainer" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.099401 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c7977bcd9-2vs8h" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.150869 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.160384 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c7977bcd9-2vs8h"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.286699 4907 scope.go:117] "RemoveContainer" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.318174 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-86cbbf5dd8-25c77"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.378841 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7dbbf4d5b8-846ft"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.467986 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fd69496d-xglm5"] Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.532520 4907 scope.go:117] "RemoveContainer" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" Mar 13 15:47:32 crc kubenswrapper[4907]: E0313 15:47:32.533015 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c\": container with ID starting with 64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c not found: ID does not exist" containerID="64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.533062 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c"} err="failed to get container status \"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c\": rpc error: code = NotFound desc = could not find container \"64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c\": container with ID starting with 64ab189d02c2d9e2e9123ec6de24b5d5f7db40d7ceaa4b16356eefe21fdc102c not found: ID does not exist" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.533088 4907 scope.go:117] "RemoveContainer" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" Mar 13 15:47:32 crc kubenswrapper[4907]: E0313 15:47:32.533414 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29\": container with ID starting with d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29 not found: ID does not exist" containerID="d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29" Mar 13 15:47:32 crc kubenswrapper[4907]: I0313 15:47:32.533450 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29"} err="failed to get container status \"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29\": rpc error: code = NotFound desc = could not find container \"d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29\": container with ID starting with d405f8c27cb2c1dbfd5ce42b31129e474a50d6e8ddffc33a12dfa64551646e29 not found: ID does not exist" Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.121113 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" event={"ID":"1d7cc945-4d68-41c8-9ea0-359af3279cc8","Type":"ContainerStarted","Data":"fea6e5248ff50df8d5cb262ab89780fb4759d674abef9321118b834ea41cbfcb"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.130566 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-86cbbf5dd8-25c77" event={"ID":"3dd72cd1-7342-40aa-8c3c-9d14375b3930","Type":"ContainerStarted","Data":"d82e65a7846c7b8476e0e378027abc6125bb877534154e1e4f897fad13bd3dd2"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.130605 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-86cbbf5dd8-25c77" event={"ID":"3dd72cd1-7342-40aa-8c3c-9d14375b3930","Type":"ContainerStarted","Data":"dacedb00970328528f11eabf2aca603e86d8a8f608fcd8855b8d709b6f89b644"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.130700 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.143288 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fd69496d-xglm5" event={"ID":"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a","Type":"ContainerStarted","Data":"2c4d11de023cb7ff9633f51fe1bcfa511e7ed99c116116d286b152f650b5ab8b"} Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.159723 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-86cbbf5dd8-25c77" podStartSLOduration=2.159703351 podStartE2EDuration="2.159703351s" podCreationTimestamp="2026-03-13 15:47:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:47:33.149610555 +0000 UTC m=+6152.049398244" watchObservedRunningTime="2026-03-13 15:47:33.159703351 +0000 UTC m=+6152.059491040" Mar 13 15:47:33 crc kubenswrapper[4907]: I0313 15:47:33.801943 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" path="/var/lib/kubelet/pods/2d9fbab8-539c-48f6-8cae-3eda23300b28/volumes" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.755706 4907 scope.go:117] "RemoveContainer" containerID="ca617bcf57734b16a051ae67c35004e1d491c4d4b7be39c47467fe81b31f47b6" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.797901 4907 scope.go:117] "RemoveContainer" containerID="29bfa2eaa41b750e3b4b9e54874f96fd9d36ce47ad55875b7c1d9bd6b63f1eef" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.871110 4907 scope.go:117] "RemoveContainer" containerID="eb5f297398bf730822a16b0adfb1835959f9d7d48e8a8a2e6d33b8a74b86db31" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.942979 4907 scope.go:117] "RemoveContainer" containerID="2d8f130696e7f9d5999a1090959bb47d9418a9bc337a9f062febc9b4224a5eb9" Mar 13 15:47:34 crc kubenswrapper[4907]: I0313 15:47:34.990189 4907 scope.go:117] "RemoveContainer" containerID="6b2aa6e8ccbdaa4de43a0268ebdb8841de04a3498e495de6ed723c183300253b" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.037006 4907 scope.go:117] "RemoveContainer" containerID="0e271cdd3f12ce6ad2a6fc124e7bdced8c6ac6c38d7c47a03deae9761bf3f57a" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.179667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fd69496d-xglm5" event={"ID":"a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a","Type":"ContainerStarted","Data":"0c5fc1a9c34f912d4a6872a0f9e97bab3fc66f28df1bf78d8be3f69d33aabd71"} Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.179751 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.191298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" event={"ID":"1d7cc945-4d68-41c8-9ea0-359af3279cc8","Type":"ContainerStarted","Data":"b32704f8f08f4c5c467b88eaaa1d87c264c5f61cec547c047a8e6beb93271ea8"} Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.192403 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.222418 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7fd69496d-xglm5" podStartSLOduration=2.537614197 podStartE2EDuration="4.222398938s" podCreationTimestamp="2026-03-13 15:47:31 +0000 UTC" firstStartedPulling="2026-03-13 15:47:32.461785113 +0000 UTC m=+6151.361572792" lastFinishedPulling="2026-03-13 15:47:34.146569844 +0000 UTC m=+6153.046357533" observedRunningTime="2026-03-13 15:47:35.204434067 +0000 UTC m=+6154.104221756" watchObservedRunningTime="2026-03-13 15:47:35.222398938 +0000 UTC m=+6154.122186647" Mar 13 15:47:35 crc kubenswrapper[4907]: I0313 15:47:35.227575 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" podStartSLOduration=2.473599566 podStartE2EDuration="4.227558458s" podCreationTimestamp="2026-03-13 15:47:31 +0000 UTC" firstStartedPulling="2026-03-13 15:47:32.389048235 +0000 UTC m=+6151.288835914" lastFinishedPulling="2026-03-13 15:47:34.143007117 +0000 UTC m=+6153.042794806" observedRunningTime="2026-03-13 15:47:35.223931149 +0000 UTC m=+6154.123718858" watchObservedRunningTime="2026-03-13 15:47:35.227558458 +0000 UTC m=+6154.127346147" Mar 13 15:47:38 crc kubenswrapper[4907]: I0313 15:47:38.370029 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:38 crc kubenswrapper[4907]: I0313 15:47:38.423974 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:38 crc kubenswrapper[4907]: I0313 15:47:38.603955 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.233452 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tthk2" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" containerID="cri-o://29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" gracePeriod=2 Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.667214 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.735874 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.823773 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") pod \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.824482 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") pod \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.824780 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") pod \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\" (UID: \"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe\") " Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.825268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities" (OuterVolumeSpecName: "utilities") pod "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" (UID: "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.825782 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.832545 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5" (OuterVolumeSpecName: "kube-api-access-6kbk5") pod "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" (UID: "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe"). InnerVolumeSpecName "kube-api-access-6kbk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.868670 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" (UID: "4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.928029 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kbk5\" (UniqueName: \"kubernetes.io/projected/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-kube-api-access-6kbk5\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:40 crc kubenswrapper[4907]: I0313 15:47:40.928165 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.246907 4907 generic.go:334] "Generic (PLEG): container finished" podID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" exitCode=0 Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.246960 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348"} Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.247036 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tthk2" event={"ID":"4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe","Type":"ContainerDied","Data":"ab17e9b088419fabf1ec951208897a9f96839403f2d2331918d09059ad41ce73"} Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.247057 4907 scope.go:117] "RemoveContainer" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.247015 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tthk2" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.276116 4907 scope.go:117] "RemoveContainer" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.282778 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.291666 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tthk2"] Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.298274 4907 scope.go:117] "RemoveContainer" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.349999 4907 scope.go:117] "RemoveContainer" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.350555 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348\": container with ID starting with 29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348 not found: ID does not exist" containerID="29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.350597 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348"} err="failed to get container status \"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348\": rpc error: code = NotFound desc = could not find container \"29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348\": container with ID starting with 29ee2de0275600572b2cd057fb10e5905301ef9ab1753a5ad562223165c95348 not found: ID does not exist" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.350623 4907 scope.go:117] "RemoveContainer" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.351264 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff\": container with ID starting with dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff not found: ID does not exist" containerID="dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.351315 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff"} err="failed to get container status \"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff\": rpc error: code = NotFound desc = could not find container \"dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff\": container with ID starting with dfea42d9a630e3e1c6c04649bc5e93bd04f889efb044725230049bd0d1de35ff not found: ID does not exist" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.351347 4907 scope.go:117] "RemoveContainer" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.351689 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916\": container with ID starting with 3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916 not found: ID does not exist" containerID="3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.351720 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916"} err="failed to get container status \"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916\": rpc error: code = NotFound desc = could not find container \"3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916\": container with ID starting with 3fc3c676404f5dc909e1ad737ea131e306c394034b644f7d1fc0ccf0fd152916 not found: ID does not exist" Mar 13 15:47:41 crc kubenswrapper[4907]: I0313 15:47:41.806241 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" path="/var/lib/kubelet/pods/4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe/volumes" Mar 13 15:47:41 crc kubenswrapper[4907]: E0313 15:47:41.888965 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb\": RecentStats: unable to find data in memory cache]" Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.241317 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6d79dff969-ssj4t" Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.351525 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.351815 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" containerID="cri-o://4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f" gracePeriod=30 Mar 13 15:47:42 crc kubenswrapper[4907]: I0313 15:47:42.352404 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" containerID="cri-o://b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931" gracePeriod=30 Mar 13 15:47:43 crc kubenswrapper[4907]: I0313 15:47:43.127301 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7fd69496d-xglm5" Mar 13 15:47:43 crc kubenswrapper[4907]: I0313 15:47:43.203360 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7dbbf4d5b8-846ft" Mar 13 15:47:46 crc kubenswrapper[4907]: I0313 15:47:46.294782 4907 generic.go:334] "Generic (PLEG): container finished" podID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerID="b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931" exitCode=0 Mar 13 15:47:46 crc kubenswrapper[4907]: I0313 15:47:46.294861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerDied","Data":"b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931"} Mar 13 15:47:47 crc kubenswrapper[4907]: I0313 15:47:47.488357 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:47:51 crc kubenswrapper[4907]: I0313 15:47:51.754354 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-86cbbf5dd8-25c77" Mar 13 15:47:52 crc kubenswrapper[4907]: E0313 15:47:52.125160 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099b49c3_666e_4bf5_9c15_32cf9c40e1e7.slice/crio-e4caf9d5f22ffb193c93c87c06b4e64f7a29f8bf0582a1a8b7f47762f5ba55eb\": RecentStats: unable to find data in memory cache]" Mar 13 15:47:57 crc kubenswrapper[4907]: I0313 15:47:57.487934 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.206019 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.206959 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-content" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.206980 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-content" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207007 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207018 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207050 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-utilities" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207063 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="extract-utilities" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207094 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207107 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" Mar 13 15:48:00 crc kubenswrapper[4907]: E0313 15:48:00.207125 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207135 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207444 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ecb3195-45a5-4ca4-b979-9a8c37ec4fbe" containerName="registry-server" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207473 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon-log" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.207488 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d9fbab8-539c-48f6-8cae-3eda23300b28" containerName="horizon" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.208572 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.211130 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.211376 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.211936 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.222265 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.270378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"auto-csr-approver-29556948-mmprl\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.373049 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"auto-csr-approver-29556948-mmprl\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.395443 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"auto-csr-approver-29556948-mmprl\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:00 crc kubenswrapper[4907]: I0313 15:48:00.539022 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:01 crc kubenswrapper[4907]: I0313 15:48:01.080129 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:48:01 crc kubenswrapper[4907]: I0313 15:48:01.083690 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:48:01 crc kubenswrapper[4907]: I0313 15:48:01.444363 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556948-mmprl" event={"ID":"673cadb2-82bd-4e26-8222-5a89c310f597","Type":"ContainerStarted","Data":"965a5ffa68a0e452b4155d7f17ef1a1f0763c8d29a45b0bbc4ebb8845a50c5f4"} Mar 13 15:48:03 crc kubenswrapper[4907]: I0313 15:48:03.464558 4907 generic.go:334] "Generic (PLEG): container finished" podID="673cadb2-82bd-4e26-8222-5a89c310f597" containerID="48d98d79386ad3e2c5200a51cdf22425a3d47af926f18ea01f01295cd4a45c13" exitCode=0 Mar 13 15:48:03 crc kubenswrapper[4907]: I0313 15:48:03.464705 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556948-mmprl" event={"ID":"673cadb2-82bd-4e26-8222-5a89c310f597","Type":"ContainerDied","Data":"48d98d79386ad3e2c5200a51cdf22425a3d47af926f18ea01f01295cd4a45c13"} Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.830766 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.865056 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") pod \"673cadb2-82bd-4e26-8222-5a89c310f597\" (UID: \"673cadb2-82bd-4e26-8222-5a89c310f597\") " Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.875581 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg" (OuterVolumeSpecName: "kube-api-access-njncg") pod "673cadb2-82bd-4e26-8222-5a89c310f597" (UID: "673cadb2-82bd-4e26-8222-5a89c310f597"). InnerVolumeSpecName "kube-api-access-njncg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:04 crc kubenswrapper[4907]: I0313 15:48:04.968916 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njncg\" (UniqueName: \"kubernetes.io/projected/673cadb2-82bd-4e26-8222-5a89c310f597-kube-api-access-njncg\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.487545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556948-mmprl" event={"ID":"673cadb2-82bd-4e26-8222-5a89c310f597","Type":"ContainerDied","Data":"965a5ffa68a0e452b4155d7f17ef1a1f0763c8d29a45b0bbc4ebb8845a50c5f4"} Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.487912 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="965a5ffa68a0e452b4155d7f17ef1a1f0763c8d29a45b0bbc4ebb8845a50c5f4" Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.487594 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556948-mmprl" Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.903688 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:48:05 crc kubenswrapper[4907]: I0313 15:48:05.913740 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556942-mdddr"] Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.487330 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-655cbf768f-dm9zz" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.156:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.156:8080: connect: connection refused" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.487910 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.554305 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2"] Mar 13 15:48:07 crc kubenswrapper[4907]: E0313 15:48:07.554747 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" containerName="oc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.554769 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" containerName="oc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.555619 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" containerName="oc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.557344 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.559432 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.565200 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2"] Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.730379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.730451 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.730510 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.798604 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed0ed6c7-e5d4-4770-a9aa-308c0853573e" path="/var/lib/kubelet/pods/ed0ed6c7-e5d4-4770-a9aa-308c0853573e/volumes" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832006 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832068 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832640 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.832684 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.856065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:07 crc kubenswrapper[4907]: I0313 15:48:07.883281 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.034211 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.046215 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.062544 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-fd12-account-create-update-kxbcs"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.074412 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kxfg2"] Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.338609 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2"] Mar 13 15:48:08 crc kubenswrapper[4907]: W0313 15:48:08.350871 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd91395e4_18e5_4da8_b482_093b71c472a3.slice/crio-6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334 WatchSource:0}: Error finding container 6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334: Status 404 returned error can't find the container with id 6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334 Mar 13 15:48:08 crc kubenswrapper[4907]: I0313 15:48:08.527178 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerStarted","Data":"6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334"} Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.169606 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.172455 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.184216 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.365573 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.365679 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.365861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.467366 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.467436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.467590 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.468111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.468192 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.488849 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"redhat-operators-7tp2h\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.500648 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.544438 4907 generic.go:334] "Generic (PLEG): container finished" podID="d91395e4-18e5-4da8-b482-093b71c472a3" containerID="2cbc12de3757ad8a4d2b65395ca25daf3aeb5685c9886ea14b7ac0bfb3d45f3b" exitCode=0 Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.544576 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"2cbc12de3757ad8a4d2b65395ca25daf3aeb5685c9886ea14b7ac0bfb3d45f3b"} Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.816969 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d6a6021-7cc4-42f9-848b-5bbfc3650ee7" path="/var/lib/kubelet/pods/0d6a6021-7cc4-42f9-848b-5bbfc3650ee7/volumes" Mar 13 15:48:09 crc kubenswrapper[4907]: I0313 15:48:09.818179 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="440d4795-da82-461c-b460-bdd5c271de4d" path="/var/lib/kubelet/pods/440d4795-da82-461c-b460-bdd5c271de4d/volumes" Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.003238 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:48:10 crc kubenswrapper[4907]: W0313 15:48:10.013276 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8c5a748_e7b7_43e0_82d3_c76673b17ea1.slice/crio-a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8 WatchSource:0}: Error finding container a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8: Status 404 returned error can't find the container with id a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8 Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.564369 4907 generic.go:334] "Generic (PLEG): container finished" podID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" exitCode=0 Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.564430 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe"} Mar 13 15:48:10 crc kubenswrapper[4907]: I0313 15:48:10.564652 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerStarted","Data":"a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8"} Mar 13 15:48:11 crc kubenswrapper[4907]: I0313 15:48:11.578979 4907 generic.go:334] "Generic (PLEG): container finished" podID="d91395e4-18e5-4da8-b482-093b71c472a3" containerID="39803b5860304bb3627a169f6cc4cff88371b582077153cce63b25eb9b7f5995" exitCode=0 Mar 13 15:48:11 crc kubenswrapper[4907]: I0313 15:48:11.579077 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"39803b5860304bb3627a169f6cc4cff88371b582077153cce63b25eb9b7f5995"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.592805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerStarted","Data":"97a66d09d1f3ef76b4030e5d8d68cadcb0f3a729374c9f8133ade6ef4e6bbf41"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.595082 4907 generic.go:334] "Generic (PLEG): container finished" podID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerID="4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f" exitCode=137 Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.595149 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerDied","Data":"4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.597664 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerStarted","Data":"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df"} Mar 13 15:48:12 crc kubenswrapper[4907]: I0313 15:48:12.636182 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" podStartSLOduration=4.37281474 podStartE2EDuration="5.636158507s" podCreationTimestamp="2026-03-13 15:48:07 +0000 UTC" firstStartedPulling="2026-03-13 15:48:09.546382159 +0000 UTC m=+6188.446169848" lastFinishedPulling="2026-03-13 15:48:10.809725926 +0000 UTC m=+6189.709513615" observedRunningTime="2026-03-13 15:48:12.611654958 +0000 UTC m=+6191.511442677" watchObservedRunningTime="2026-03-13 15:48:12.636158507 +0000 UTC m=+6191.535946196" Mar 13 15:48:12 crc kubenswrapper[4907]: E0313 15:48:12.651503 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafd4001d_40a5_4a09_ab05_36cbd4d06eb4.slice/crio-conmon-4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.321406 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.351629 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.394625 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts" (OuterVolumeSpecName: "scripts") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454030 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454207 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454301 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454366 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") pod \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\" (UID: \"afd4001d-40a5-4a09-ab05-36cbd4d06eb4\") " Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.454792 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs" (OuterVolumeSpecName: "logs") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.455069 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-logs\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.455095 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.458036 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.458108 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh" (OuterVolumeSpecName: "kube-api-access-ml6dh") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "kube-api-access-ml6dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.481348 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data" (OuterVolumeSpecName: "config-data") pod "afd4001d-40a5-4a09-ab05-36cbd4d06eb4" (UID: "afd4001d-40a5-4a09-ab05-36cbd4d06eb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.556677 4907 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.556729 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.556750 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml6dh\" (UniqueName: \"kubernetes.io/projected/afd4001d-40a5-4a09-ab05-36cbd4d06eb4-kube-api-access-ml6dh\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.628775 4907 generic.go:334] "Generic (PLEG): container finished" podID="d91395e4-18e5-4da8-b482-093b71c472a3" containerID="97a66d09d1f3ef76b4030e5d8d68cadcb0f3a729374c9f8133ade6ef4e6bbf41" exitCode=0 Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.628850 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"97a66d09d1f3ef76b4030e5d8d68cadcb0f3a729374c9f8133ade6ef4e6bbf41"} Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.630544 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-655cbf768f-dm9zz" event={"ID":"afd4001d-40a5-4a09-ab05-36cbd4d06eb4","Type":"ContainerDied","Data":"8eed0ad989f3d5a163602bd4e967ac325bbb242d625bd6a2d8dac4fb6b57f365"} Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.630588 4907 scope.go:117] "RemoveContainer" containerID="b42f8e33dfb6318b768e8c6b32af48621974ee97af5dab3f3e4bc582bbfc2931" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.630634 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-655cbf768f-dm9zz" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.689398 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.700511 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-655cbf768f-dm9zz"] Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.792569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" path="/var/lib/kubelet/pods/afd4001d-40a5-4a09-ab05-36cbd4d06eb4/volumes" Mar 13 15:48:13 crc kubenswrapper[4907]: I0313 15:48:13.814920 4907 scope.go:117] "RemoveContainer" containerID="4d9d087b4ad0675d46c48144b039251dc336275fe3320abbe51816f548542e9f" Mar 13 15:48:14 crc kubenswrapper[4907]: I0313 15:48:14.641361 4907 generic.go:334] "Generic (PLEG): container finished" podID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" exitCode=0 Mar 13 15:48:14 crc kubenswrapper[4907]: I0313 15:48:14.641434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df"} Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.057134 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.102681 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") pod \"d91395e4-18e5-4da8-b482-093b71c472a3\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.102837 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") pod \"d91395e4-18e5-4da8-b482-093b71c472a3\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.102891 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") pod \"d91395e4-18e5-4da8-b482-093b71c472a3\" (UID: \"d91395e4-18e5-4da8-b482-093b71c472a3\") " Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.104797 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle" (OuterVolumeSpecName: "bundle") pod "d91395e4-18e5-4da8-b482-093b71c472a3" (UID: "d91395e4-18e5-4da8-b482-093b71c472a3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.108851 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv" (OuterVolumeSpecName: "kube-api-access-sd9hv") pod "d91395e4-18e5-4da8-b482-093b71c472a3" (UID: "d91395e4-18e5-4da8-b482-093b71c472a3"). InnerVolumeSpecName "kube-api-access-sd9hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.111571 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util" (OuterVolumeSpecName: "util") pod "d91395e4-18e5-4da8-b482-093b71c472a3" (UID: "d91395e4-18e5-4da8-b482-093b71c472a3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.207407 4907 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.207730 4907 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d91395e4-18e5-4da8-b482-093b71c472a3-util\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.207746 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd9hv\" (UniqueName: \"kubernetes.io/projected/d91395e4-18e5-4da8-b482-093b71c472a3-kube-api-access-sd9hv\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.661533 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" event={"ID":"d91395e4-18e5-4da8-b482-093b71c472a3","Type":"ContainerDied","Data":"6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334"} Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.661772 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e048b993c2bad9cfa6a447e0cee7c6ba7ef45bc00d56e28386ad86b9c523334" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.661837 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2" Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.666832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerStarted","Data":"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333"} Mar 13 15:48:15 crc kubenswrapper[4907]: I0313 15:48:15.687623 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7tp2h" podStartSLOduration=2.392888227 podStartE2EDuration="6.687607617s" podCreationTimestamp="2026-03-13 15:48:09 +0000 UTC" firstStartedPulling="2026-03-13 15:48:10.747611058 +0000 UTC m=+6189.647398757" lastFinishedPulling="2026-03-13 15:48:15.042330458 +0000 UTC m=+6193.942118147" observedRunningTime="2026-03-13 15:48:15.682635822 +0000 UTC m=+6194.582423531" watchObservedRunningTime="2026-03-13 15:48:15.687607617 +0000 UTC m=+6194.587395306" Mar 13 15:48:17 crc kubenswrapper[4907]: I0313 15:48:17.049872 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:48:17 crc kubenswrapper[4907]: I0313 15:48:17.059023 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tlb57"] Mar 13 15:48:17 crc kubenswrapper[4907]: I0313 15:48:17.795307 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c6b679-96d2-407c-8b94-cd908d61945b" path="/var/lib/kubelet/pods/13c6b679-96d2-407c-8b94-cd908d61945b/volumes" Mar 13 15:48:19 crc kubenswrapper[4907]: I0313 15:48:19.501257 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:19 crc kubenswrapper[4907]: I0313 15:48:19.501551 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:48:20 crc kubenswrapper[4907]: I0313 15:48:20.678368 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:20 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:20 crc kubenswrapper[4907]: > Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.303677 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb"] Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.304633 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="pull" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.304649 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="pull" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.304665 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.304671 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.310452 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="util" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.310522 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="util" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.310564 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="extract" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.310574 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="extract" Mar 13 15:48:24 crc kubenswrapper[4907]: E0313 15:48:24.310607 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.310615 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311044 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311063 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d91395e4-18e5-4da8-b482-093b71c472a3" containerName="extract" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311083 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd4001d-40a5-4a09-ab05-36cbd4d06eb4" containerName="horizon-log" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.311813 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.316519 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.316763 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-kkxtc" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.316931 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.330185 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.415353 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n5jn\" (UniqueName: \"kubernetes.io/projected/92caa667-d9ee-4043-95b1-6475bcc7439b-kube-api-access-9n5jn\") pod \"obo-prometheus-operator-68bc856cb9-w2ngb\" (UID: \"92caa667-d9ee-4043-95b1-6475bcc7439b\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.434898 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.436228 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.440193 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.440546 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-jm8k4" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.458238 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.471593 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.473336 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.506141 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.518077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n5jn\" (UniqueName: \"kubernetes.io/projected/92caa667-d9ee-4043-95b1-6475bcc7439b-kube-api-access-9n5jn\") pod \"obo-prometheus-operator-68bc856cb9-w2ngb\" (UID: \"92caa667-d9ee-4043-95b1-6475bcc7439b\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.546427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n5jn\" (UniqueName: \"kubernetes.io/projected/92caa667-d9ee-4043-95b1-6475bcc7439b-kube-api-access-9n5jn\") pod \"obo-prometheus-operator-68bc856cb9-w2ngb\" (UID: \"92caa667-d9ee-4043-95b1-6475bcc7439b\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620135 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620271 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620304 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.620344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.639348 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.650081 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-9hfqr"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.670616 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.681150 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-xctpc" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.682787 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.709638 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-9hfqr"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.726742 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.727123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.727215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.727325 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.731681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.733446 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d188fa9a-103d-49af-9626-90b7a3a5d70c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq\" (UID: \"d188fa9a-103d-49af-9626-90b7a3a5d70c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.733831 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.759360 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.765138 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/683c61a5-c39a-400b-b673-19bbaf286482-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz\" (UID: \"683c61a5-c39a-400b-b673-19bbaf286482\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.801580 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.830533 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqnp6\" (UniqueName: \"kubernetes.io/projected/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-kube-api-access-fqnp6\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.830708 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.878985 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-2n9b9"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.881035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.883716 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-6699h" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.893446 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-2n9b9"] Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.935213 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqnp6\" (UniqueName: \"kubernetes.io/projected/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-kube-api-access-fqnp6\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.935477 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.943742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:24 crc kubenswrapper[4907]: I0313 15:48:24.973489 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqnp6\" (UniqueName: \"kubernetes.io/projected/dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d-kube-api-access-fqnp6\") pod \"observability-operator-59bdc8b94-9hfqr\" (UID: \"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d\") " pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.036904 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm24m\" (UniqueName: \"kubernetes.io/projected/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-kube-api-access-nm24m\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.036965 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-openshift-service-ca\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.139934 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm24m\" (UniqueName: \"kubernetes.io/projected/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-kube-api-access-nm24m\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.140017 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-openshift-service-ca\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.141173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-openshift-service-ca\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.165745 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm24m\" (UniqueName: \"kubernetes.io/projected/c5a33f26-edc6-4cb0-b9d8-58ee6113fecc-kube-api-access-nm24m\") pod \"perses-operator-5bf474d74f-2n9b9\" (UID: \"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc\") " pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.239337 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.361038 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.436311 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb"] Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.554106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq"] Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.654680 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz"] Mar 13 15:48:25 crc kubenswrapper[4907]: W0313 15:48:25.668052 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod683c61a5_c39a_400b_b673_19bbaf286482.slice/crio-c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699 WatchSource:0}: Error finding container c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699: Status 404 returned error can't find the container with id c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699 Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.897282 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-9hfqr"] Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.914089 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" event={"ID":"d188fa9a-103d-49af-9626-90b7a3a5d70c","Type":"ContainerStarted","Data":"f0d8af2bb6aee7cb78c99769c70b55465d6cb45654b154e2d62291621b414e7d"} Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.925504 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" event={"ID":"683c61a5-c39a-400b-b673-19bbaf286482","Type":"ContainerStarted","Data":"c888be885a02727fc3c9a0bf84f48814d7f5af0fa589398b52218f373a6b1699"} Mar 13 15:48:25 crc kubenswrapper[4907]: I0313 15:48:25.938053 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" event={"ID":"92caa667-d9ee-4043-95b1-6475bcc7439b","Type":"ContainerStarted","Data":"92c25da4a3fec1bfea2fada32dd1fb03ce8186eae22530574b27ef9d987a11a4"} Mar 13 15:48:26 crc kubenswrapper[4907]: I0313 15:48:26.025491 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-2n9b9"] Mar 13 15:48:26 crc kubenswrapper[4907]: I0313 15:48:26.952897 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" event={"ID":"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc","Type":"ContainerStarted","Data":"33323ecd7b6da6cb3e421e9ee7b60e850ef23ab3122705fdfede0831368e2243"} Mar 13 15:48:26 crc kubenswrapper[4907]: I0313 15:48:26.957953 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" event={"ID":"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d","Type":"ContainerStarted","Data":"251bf2b12c5832c0144196c77dc1bb3b002c61ccb327435531f68fbea3f397fe"} Mar 13 15:48:30 crc kubenswrapper[4907]: I0313 15:48:30.566405 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:30 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:30 crc kubenswrapper[4907]: > Mar 13 15:48:35 crc kubenswrapper[4907]: I0313 15:48:35.323762 4907 scope.go:117] "RemoveContainer" containerID="7414c52bdeec9a9b8ecea00c6ccd0ca4cd0ecf4d6c333567e9430b7214c0182e" Mar 13 15:48:37 crc kubenswrapper[4907]: I0313 15:48:37.120158 4907 scope.go:117] "RemoveContainer" containerID="75dd2c3ec3d1e72efcbe86eb956d5d57de584abd904ef3f55fe25aad75bb5498" Mar 13 15:48:37 crc kubenswrapper[4907]: I0313 15:48:37.230541 4907 scope.go:117] "RemoveContainer" containerID="7c596675c8272c2dfe2507df7b83ab578498cef4ae7b5f27995638598d3ea046" Mar 13 15:48:37 crc kubenswrapper[4907]: I0313 15:48:37.396430 4907 scope.go:117] "RemoveContainer" containerID="38713faad1d72bf574ce7d6ae2bd740312997c935666c1fb81c567bd339dc0d0" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.116688 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" event={"ID":"dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d","Type":"ContainerStarted","Data":"44b0e8afc6cb817bd60894134bccaf31cfd9deaca012003d688c414c82d33f25"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.117340 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.121958 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" event={"ID":"92caa667-d9ee-4043-95b1-6475bcc7439b","Type":"ContainerStarted","Data":"2639c49c39197ee0e143615b69caabdfef7076a9f335b899f297bd55a60ed0bb"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.123147 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.136441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" event={"ID":"c5a33f26-edc6-4cb0-b9d8-58ee6113fecc","Type":"ContainerStarted","Data":"fe3213f56f2b0e957e334972f5b7c9ae49f4208945edd4c62c8b0b5ee34da857"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.136988 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.143792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" event={"ID":"d188fa9a-103d-49af-9626-90b7a3a5d70c","Type":"ContainerStarted","Data":"e9139275039d32450405a70812febf14bac9d1aa924b322f48f5b14bc30cc0a8"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.149529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" event={"ID":"683c61a5-c39a-400b-b673-19bbaf286482","Type":"ContainerStarted","Data":"0b60abf6ed19147654aff611bdc1b1ae57ef3fc7348dd8abd597b16cee46e52f"} Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.163266 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-9hfqr" podStartSLOduration=2.828113116 podStartE2EDuration="14.163247371s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.895928557 +0000 UTC m=+6204.795716236" lastFinishedPulling="2026-03-13 15:48:37.231062802 +0000 UTC m=+6216.130850491" observedRunningTime="2026-03-13 15:48:38.140343876 +0000 UTC m=+6217.040131585" watchObservedRunningTime="2026-03-13 15:48:38.163247371 +0000 UTC m=+6217.063035060" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.197381 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" podStartSLOduration=3.117126732 podStartE2EDuration="14.197361733s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:26.041064643 +0000 UTC m=+6204.940852332" lastFinishedPulling="2026-03-13 15:48:37.121299644 +0000 UTC m=+6216.021087333" observedRunningTime="2026-03-13 15:48:38.193316703 +0000 UTC m=+6217.093104392" watchObservedRunningTime="2026-03-13 15:48:38.197361733 +0000 UTC m=+6217.097149422" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.240425 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-w2ngb" podStartSLOduration=2.546116391 podStartE2EDuration="14.240400789s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.427457188 +0000 UTC m=+6204.327244877" lastFinishedPulling="2026-03-13 15:48:37.121741586 +0000 UTC m=+6216.021529275" observedRunningTime="2026-03-13 15:48:38.236752289 +0000 UTC m=+6217.136539998" watchObservedRunningTime="2026-03-13 15:48:38.240400789 +0000 UTC m=+6217.140188478" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.265429 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz" podStartSLOduration=2.842234182 podStartE2EDuration="14.265406803s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.697029523 +0000 UTC m=+6204.596817212" lastFinishedPulling="2026-03-13 15:48:37.120202144 +0000 UTC m=+6216.019989833" observedRunningTime="2026-03-13 15:48:38.260411946 +0000 UTC m=+6217.160199655" watchObservedRunningTime="2026-03-13 15:48:38.265406803 +0000 UTC m=+6217.165194492" Mar 13 15:48:38 crc kubenswrapper[4907]: I0313 15:48:38.318086 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq" podStartSLOduration=2.774009668 podStartE2EDuration="14.318061711s" podCreationTimestamp="2026-03-13 15:48:24 +0000 UTC" firstStartedPulling="2026-03-13 15:48:25.583130921 +0000 UTC m=+6204.482918610" lastFinishedPulling="2026-03-13 15:48:37.127182964 +0000 UTC m=+6216.026970653" observedRunningTime="2026-03-13 15:48:38.305170589 +0000 UTC m=+6217.204958288" watchObservedRunningTime="2026-03-13 15:48:38.318061711 +0000 UTC m=+6217.217849400" Mar 13 15:48:40 crc kubenswrapper[4907]: I0313 15:48:40.563556 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:40 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:40 crc kubenswrapper[4907]: > Mar 13 15:48:45 crc kubenswrapper[4907]: I0313 15:48:45.364874 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-2n9b9" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.041727 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.042270 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.334851 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.335136 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" containerID="cri-o://e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" gracePeriod=2 Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.349469 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.396660 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.397240 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.397265 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.397495 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerName="openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.398447 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.407967 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.428503 4907 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc7436bb-91c1-4364-9ef4-dba14540c5c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T15:48:48Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:d765b589a5f7bc8573b3b192ed265654699012e6342cc4829bd8ea65a7b239a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jj76j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T15:48:48Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.429478 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.430141 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-jj76j openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.444444 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.480548 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.482445 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.503510 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.503592 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.503747 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.510297 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.525204 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.539125 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="74a5c619-2dde-4f75-9106-cad6a39265ce" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.605896 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606234 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606312 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606343 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxg26\" (UniqueName: \"kubernetes.io/projected/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-kube-api-access-vxg26\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.606467 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.607589 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.609391 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.610868 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.620795 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bl4lv" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.620930 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.621116 4907 projected.go:194] Error preparing data for projected volume kube-api-access-jj76j for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:48 crc kubenswrapper[4907]: E0313 15:48:48.621176 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j podName:bc7436bb-91c1-4364-9ef4-dba14540c5c2 nodeName:}" failed. No retries permitted until 2026-03-13 15:48:49.121156849 +0000 UTC m=+6228.020944538 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-jj76j" (UniqueName: "kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j") pod "openstackclient" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.639221 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4tqw\" (UniqueName: \"kubernetes.io/projected/53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592-kube-api-access-m4tqw\") pod \"kube-state-metrics-0\" (UID: \"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592\") " pod="openstack/kube-state-metrics-0" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.709765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxg26\" (UniqueName: \"kubernetes.io/projected/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-kube-api-access-vxg26\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.714794 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config-secret\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.714837 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-openstack-config\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.821261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxg26\" (UniqueName: \"kubernetes.io/projected/2316d199-eef1-4ac1-9f63-cc73cba8d4c7-kube-api-access-vxg26\") pod \"openstackclient\" (UID: \"2316d199-eef1-4ac1-9f63-cc73cba8d4c7\") " pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.823102 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.868464 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4tqw\" (UniqueName: \"kubernetes.io/projected/53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592-kube-api-access-m4tqw\") pod \"kube-state-metrics-0\" (UID: \"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592\") " pod="openstack/kube-state-metrics-0" Mar 13 15:48:48 crc kubenswrapper[4907]: I0313 15:48:48.910715 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4tqw\" (UniqueName: \"kubernetes.io/projected/53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592-kube-api-access-m4tqw\") pod \"kube-state-metrics-0\" (UID: \"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592\") " pod="openstack/kube-state-metrics-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.006399 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.184211 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") pod \"openstackclient\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " pod="openstack/openstackclient" Mar 13 15:48:49 crc kubenswrapper[4907]: E0313 15:48:49.186694 4907 projected.go:194] Error preparing data for projected volume kube-api-access-jj76j for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:49 crc kubenswrapper[4907]: E0313 15:48:49.186758 4907 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j podName:bc7436bb-91c1-4364-9ef4-dba14540c5c2 nodeName:}" failed. No retries permitted until 2026-03-13 15:48:50.186739182 +0000 UTC m=+6229.086526861 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-jj76j" (UniqueName: "kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j") pod "openstackclient" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (bc7436bb-91c1-4364-9ef4-dba14540c5c2) does not match the UID in record. The object might have been deleted and then recreated Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.267816 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.271804 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.295183 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.300039 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.388471 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.390973 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409502 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-hc86w" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409673 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409793 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.409907 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.410029 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.423600 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.490771 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") pod \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.490864 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") pod \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\" (UID: \"bc7436bb-91c1-4364-9ef4-dba14540c5c2\") " Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.492434 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bc7436bb-91c1-4364-9ef4-dba14540c5c2" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497465 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497538 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhctv\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-kube-api-access-mhctv\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497776 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497818 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497845 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.497896 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.498183 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.498406 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj76j\" (UniqueName: \"kubernetes.io/projected/bc7436bb-91c1-4364-9ef4-dba14540c5c2-kube-api-access-jj76j\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.498426 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.512270 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bc7436bb-91c1-4364-9ef4-dba14540c5c2" (UID: "bc7436bb-91c1-4364-9ef4-dba14540c5c2"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602277 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhctv\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-kube-api-access-mhctv\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602491 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602529 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602546 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.602605 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bc7436bb-91c1-4364-9ef4-dba14540c5c2-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.613137 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.619220 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.623238 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d29f6de0-3229-477c-86a6-550f2d465592-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.623687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.636944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.641861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d29f6de0-3229-477c-86a6-550f2d465592-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.647788 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.718305 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhctv\" (UniqueName: \"kubernetes.io/projected/d29f6de0-3229-477c-86a6-550f2d465592-kube-api-access-mhctv\") pod \"alertmanager-metric-storage-0\" (UID: \"d29f6de0-3229-477c-86a6-550f2d465592\") " pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.821399 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" path="/var/lib/kubelet/pods/bc7436bb-91c1-4364-9ef4-dba14540c5c2/volumes" Mar 13 15:48:49 crc kubenswrapper[4907]: I0313 15:48:49.925242 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.002142 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.004724 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.019520 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.028008 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.028807 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.028947 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-5kg6f" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029042 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029140 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029232 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.029325 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.040283 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.079597 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.117397 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122706 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qphj\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-kube-api-access-9qphj\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122768 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122806 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.122856 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.135969 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136105 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136188 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136312 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e5a8905-7f71-4979-aee3-f81e772ea46a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136400 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.136447 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.152944 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.180961 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5svds"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.201974 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c036-account-create-update-26jl8"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238097 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qphj\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-kube-api-access-9qphj\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238157 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238202 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238238 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238279 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238304 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238333 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238370 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e5a8905-7f71-4979-aee3-f81e772ea46a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238396 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.238418 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.242341 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.242805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.243279 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3e5a8905-7f71-4979-aee3-f81e772ea46a-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.253935 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.258292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3e5a8905-7f71-4979-aee3-f81e772ea46a-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.260720 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.267715 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.267955 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3e5a8905-7f71-4979-aee3-f81e772ea46a-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.273612 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qphj\" (UniqueName: \"kubernetes.io/projected/3e5a8905-7f71-4979-aee3-f81e772ea46a-kube-api-access-9qphj\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.312611 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2316d199-eef1-4ac1-9f63-cc73cba8d4c7","Type":"ContainerStarted","Data":"e822b148bbadd75b0f6facc5f6ac13f0b1cab8ff45ef96f5417ba6cb764deb8a"} Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.313864 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.314585 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592","Type":"ContainerStarted","Data":"2356f8f220308508437db8c12932926e505f0ad6e95771702f34ffa0f6d6e8e1"} Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.324232 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bc7436bb-91c1-4364-9ef4-dba14540c5c2" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.379976 4907 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.380015 4907 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/18cb5e38569b223e596eb1f57078b15c9d63dca5ccb98a8253238ad08ba4dd89/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.534531 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac09aaac-45f0-4ecf-8f08-3e715b4f4606\") pod \"prometheus-metric-storage-0\" (UID: \"3e5a8905-7f71-4979-aee3-f81e772ea46a\") " pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.594596 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:48:50 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:48:50 crc kubenswrapper[4907]: > Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.678046 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.915276 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Mar 13 15:48:50 crc kubenswrapper[4907]: I0313 15:48:50.929776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.063068 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") pod \"74a5c619-2dde-4f75-9106-cad6a39265ce\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.063930 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") pod \"74a5c619-2dde-4f75-9106-cad6a39265ce\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.064165 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") pod \"74a5c619-2dde-4f75-9106-cad6a39265ce\" (UID: \"74a5c619-2dde-4f75-9106-cad6a39265ce\") " Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.072321 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz" (OuterVolumeSpecName: "kube-api-access-85lwz") pod "74a5c619-2dde-4f75-9106-cad6a39265ce" (UID: "74a5c619-2dde-4f75-9106-cad6a39265ce"). InnerVolumeSpecName "kube-api-access-85lwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.133787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "74a5c619-2dde-4f75-9106-cad6a39265ce" (UID: "74a5c619-2dde-4f75-9106-cad6a39265ce"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.146285 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "74a5c619-2dde-4f75-9106-cad6a39265ce" (UID: "74a5c619-2dde-4f75-9106-cad6a39265ce"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.168377 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.168413 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85lwz\" (UniqueName: \"kubernetes.io/projected/74a5c619-2dde-4f75-9106-cad6a39265ce-kube-api-access-85lwz\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.168423 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/74a5c619-2dde-4f75-9106-cad6a39265ce-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.325383 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.328781 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"3e839fa1cb364797ef12f2e906290509ded0419ffb1acfd515495d8304c579b8"} Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.330814 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2316d199-eef1-4ac1-9f63-cc73cba8d4c7","Type":"ContainerStarted","Data":"8caed9aacdfdeb98c939313e87104e0a2c29efecca834afa9daa99b5055ebc8e"} Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.337195 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592","Type":"ContainerStarted","Data":"77ad5fed4d3fec5023dd6cfa95b8071d1e0afddd56cc6b479d88fc6dfcc12714"} Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.337610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.339279 4907 generic.go:334] "Generic (PLEG): container finished" podID="74a5c619-2dde-4f75-9106-cad6a39265ce" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" exitCode=137 Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.339320 4907 scope.go:117] "RemoveContainer" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.339436 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.351988 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.351964249 podStartE2EDuration="3.351964249s" podCreationTimestamp="2026-03-13 15:48:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:48:51.349527503 +0000 UTC m=+6230.249315192" watchObservedRunningTime="2026-03-13 15:48:51.351964249 +0000 UTC m=+6230.251751958" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.383617 4907 scope.go:117] "RemoveContainer" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" Mar 13 15:48:51 crc kubenswrapper[4907]: E0313 15:48:51.385076 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff\": container with ID starting with e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff not found: ID does not exist" containerID="e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.385137 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff"} err="failed to get container status \"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff\": rpc error: code = NotFound desc = could not find container \"e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff\": container with ID starting with e99b6a0dae893428515991f160f7db931ea2980ec0fe9b55c8e67e61a9c011ff not found: ID does not exist" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.386009 4907 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="74a5c619-2dde-4f75-9106-cad6a39265ce" podUID="2316d199-eef1-4ac1-9f63-cc73cba8d4c7" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.390022 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.95926718 podStartE2EDuration="3.390006349s" podCreationTimestamp="2026-03-13 15:48:48 +0000 UTC" firstStartedPulling="2026-03-13 15:48:49.955482355 +0000 UTC m=+6228.855270044" lastFinishedPulling="2026-03-13 15:48:50.386221524 +0000 UTC m=+6229.286009213" observedRunningTime="2026-03-13 15:48:51.382323039 +0000 UTC m=+6230.282110728" watchObservedRunningTime="2026-03-13 15:48:51.390006349 +0000 UTC m=+6230.289794038" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.794211 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d" path="/var/lib/kubelet/pods/0e98ef6f-8765-4e4d-ab91-7dcdcb3f788d/volumes" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.795464 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a5c619-2dde-4f75-9106-cad6a39265ce" path="/var/lib/kubelet/pods/74a5c619-2dde-4f75-9106-cad6a39265ce/volumes" Mar 13 15:48:51 crc kubenswrapper[4907]: I0313 15:48:51.796156 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="978c19e3-b11d-44ef-86d5-6b0674e488d1" path="/var/lib/kubelet/pods/978c19e3-b11d-44ef-86d5-6b0674e488d1/volumes" Mar 13 15:48:52 crc kubenswrapper[4907]: I0313 15:48:52.353450 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"f98c017c2a67861fb920f9dadf2d97df3ff6075477cb06f3b46a33ec5ad21c7b"} Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.033519 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.044637 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-z7wl4"] Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.413264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"998196e6dbaae3637cbc8ca152cc5014c622bb70e9857a79cc0522fca4602736"} Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.418986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"a3daab1d7da324d8c75c9e53d1520c30cb91a375d6d03e39cbc921749c0bb23d"} Mar 13 15:48:57 crc kubenswrapper[4907]: I0313 15:48:57.801397 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a43a8ffb-4f15-4524-b750-517442c1f561" path="/var/lib/kubelet/pods/a43a8ffb-4f15-4524-b750-517442c1f561/volumes" Mar 13 15:48:59 crc kubenswrapper[4907]: I0313 15:48:59.011519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 15:49:00 crc kubenswrapper[4907]: I0313 15:49:00.557550 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" probeResult="failure" output=< Mar 13 15:49:00 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:49:00 crc kubenswrapper[4907]: > Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.473864 4907 generic.go:334] "Generic (PLEG): container finished" podID="3e5a8905-7f71-4979-aee3-f81e772ea46a" containerID="a3daab1d7da324d8c75c9e53d1520c30cb91a375d6d03e39cbc921749c0bb23d" exitCode=0 Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.474021 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerDied","Data":"a3daab1d7da324d8c75c9e53d1520c30cb91a375d6d03e39cbc921749c0bb23d"} Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.476076 4907 generic.go:334] "Generic (PLEG): container finished" podID="d29f6de0-3229-477c-86a6-550f2d465592" containerID="998196e6dbaae3637cbc8ca152cc5014c622bb70e9857a79cc0522fca4602736" exitCode=0 Mar 13 15:49:03 crc kubenswrapper[4907]: I0313 15:49:03.476115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerDied","Data":"998196e6dbaae3637cbc8ca152cc5014c622bb70e9857a79cc0522fca4602736"} Mar 13 15:49:06 crc kubenswrapper[4907]: I0313 15:49:06.515044 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"ed52a0746ec744f915cfe385f5b3d69db47e9fac5239ec3264f15b6665ddca8f"} Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.558050 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.566911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"d29f6de0-3229-477c-86a6-550f2d465592","Type":"ContainerStarted","Data":"494105b0e22ae748f4d76099d291149db9c96acc20858558d71b2c1c8372bb0c"} Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.567295 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.570986 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.611648 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=5.368136396 podStartE2EDuration="20.611629335s" podCreationTimestamp="2026-03-13 15:48:49 +0000 UTC" firstStartedPulling="2026-03-13 15:48:50.977982331 +0000 UTC m=+6229.877770020" lastFinishedPulling="2026-03-13 15:49:06.22147526 +0000 UTC m=+6245.121262959" observedRunningTime="2026-03-13 15:49:09.595809172 +0000 UTC m=+6248.495596861" watchObservedRunningTime="2026-03-13 15:49:09.611629335 +0000 UTC m=+6248.511417024" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.640482 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:09 crc kubenswrapper[4907]: I0313 15:49:09.813317 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:49:11 crc kubenswrapper[4907]: I0313 15:49:11.586732 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"38f43333273c3672449a529b41205de80c2ce59a4d18582d8c5417bdd810b24a"} Mar 13 15:49:11 crc kubenswrapper[4907]: I0313 15:49:11.586919 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7tp2h" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" containerID="cri-o://49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" gracePeriod=2 Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.061068 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.146805 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") pod \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.147106 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") pod \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.147435 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") pod \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\" (UID: \"e8c5a748-e7b7-43e0-82d3-c76673b17ea1\") " Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.147763 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities" (OuterVolumeSpecName: "utilities") pod "e8c5a748-e7b7-43e0-82d3-c76673b17ea1" (UID: "e8c5a748-e7b7-43e0-82d3-c76673b17ea1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.148230 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.152773 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq" (OuterVolumeSpecName: "kube-api-access-ppphq") pod "e8c5a748-e7b7-43e0-82d3-c76673b17ea1" (UID: "e8c5a748-e7b7-43e0-82d3-c76673b17ea1"). InnerVolumeSpecName "kube-api-access-ppphq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.250783 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppphq\" (UniqueName: \"kubernetes.io/projected/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-kube-api-access-ppphq\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.277750 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8c5a748-e7b7-43e0-82d3-c76673b17ea1" (UID: "e8c5a748-e7b7-43e0-82d3-c76673b17ea1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.352531 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8c5a748-e7b7-43e0-82d3-c76673b17ea1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.606985 4907 generic.go:334] "Generic (PLEG): container finished" podID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" exitCode=0 Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333"} Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607080 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tp2h" event={"ID":"e8c5a748-e7b7-43e0-82d3-c76673b17ea1","Type":"ContainerDied","Data":"a029d38e007ef1dd2747ad9ff6b5a9cef6fcfa563fa0dc97f6df3132418979f8"} Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607100 4907 scope.go:117] "RemoveContainer" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.607137 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tp2h" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.632055 4907 scope.go:117] "RemoveContainer" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.646492 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.658968 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7tp2h"] Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.687673 4907 scope.go:117] "RemoveContainer" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.709559 4907 scope.go:117] "RemoveContainer" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" Mar 13 15:49:12 crc kubenswrapper[4907]: E0313 15:49:12.710174 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333\": container with ID starting with 49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333 not found: ID does not exist" containerID="49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.710283 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333"} err="failed to get container status \"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333\": rpc error: code = NotFound desc = could not find container \"49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333\": container with ID starting with 49ef110711e5f738221014e1ce44cead4c2950560589e797e31382bd54cac333 not found: ID does not exist" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.710358 4907 scope.go:117] "RemoveContainer" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" Mar 13 15:49:12 crc kubenswrapper[4907]: E0313 15:49:12.710853 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df\": container with ID starting with 2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df not found: ID does not exist" containerID="2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.710981 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df"} err="failed to get container status \"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df\": rpc error: code = NotFound desc = could not find container \"2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df\": container with ID starting with 2d7e94bcae9c69807004704893b2f10f4c520e505aa6a59bbe61e348ddae96df not found: ID does not exist" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.711011 4907 scope.go:117] "RemoveContainer" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" Mar 13 15:49:12 crc kubenswrapper[4907]: E0313 15:49:12.711498 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe\": container with ID starting with d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe not found: ID does not exist" containerID="d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe" Mar 13 15:49:12 crc kubenswrapper[4907]: I0313 15:49:12.711569 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe"} err="failed to get container status \"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe\": rpc error: code = NotFound desc = could not find container \"d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe\": container with ID starting with d49b21ae936fc12c5a09e43efc846d4e58d2f3ad32465957c3a2c3b2abeb9cfe not found: ID does not exist" Mar 13 15:49:13 crc kubenswrapper[4907]: I0313 15:49:13.797679 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" path="/var/lib/kubelet/pods/e8c5a748-e7b7-43e0-82d3-c76673b17ea1/volumes" Mar 13 15:49:15 crc kubenswrapper[4907]: I0313 15:49:15.645168 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"a7f3d5e1b9c30bf5160801062d9ffb06f51eec5b64919c41a4a982e147424b23"} Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.041796 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.042157 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.675383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3e5a8905-7f71-4979-aee3-f81e772ea46a","Type":"ContainerStarted","Data":"22eedd14ce51e6e1c8b08afa6ad93b877c9db3053519763c8f6b052b58e2dddc"} Mar 13 15:49:18 crc kubenswrapper[4907]: I0313 15:49:18.707169 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=4.163441901 podStartE2EDuration="30.70714764s" podCreationTimestamp="2026-03-13 15:48:48 +0000 UTC" firstStartedPulling="2026-03-13 15:48:51.324819928 +0000 UTC m=+6230.224607617" lastFinishedPulling="2026-03-13 15:49:17.868525667 +0000 UTC m=+6256.768313356" observedRunningTime="2026-03-13 15:49:18.7009332 +0000 UTC m=+6257.600720899" watchObservedRunningTime="2026-03-13 15:49:18.70714764 +0000 UTC m=+6257.606935329" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.678490 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.679091 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.680831 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:20 crc kubenswrapper[4907]: I0313 15:49:20.701466 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.194800 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:22 crc kubenswrapper[4907]: E0313 15:49:22.195626 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195646 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" Mar 13 15:49:22 crc kubenswrapper[4907]: E0313 15:49:22.195671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-utilities" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195679 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-utilities" Mar 13 15:49:22 crc kubenswrapper[4907]: E0313 15:49:22.195710 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-content" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195719 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="extract-content" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.195997 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8c5a748-e7b7-43e0-82d3-c76673b17ea1" containerName="registry-server" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.198463 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.204686 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.205051 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.208126 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260458 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260536 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260627 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.260840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.261167 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.261221 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.261307 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362812 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362933 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362958 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.362994 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363021 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363044 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363096 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363467 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.363571 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.368913 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.370594 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.377584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.377639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.383635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"ceilometer-0\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " pod="openstack/ceilometer-0" Mar 13 15:49:22 crc kubenswrapper[4907]: I0313 15:49:22.536622 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:23 crc kubenswrapper[4907]: I0313 15:49:23.020620 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:23 crc kubenswrapper[4907]: I0313 15:49:23.741084 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"ae9b3358826e66742e51bbbc204d8b30a566e7e0b936fca1df7ae895ee7043e9"} Mar 13 15:49:24 crc kubenswrapper[4907]: I0313 15:49:24.759862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0"} Mar 13 15:49:24 crc kubenswrapper[4907]: I0313 15:49:24.760517 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0"} Mar 13 15:49:25 crc kubenswrapper[4907]: I0313 15:49:25.771419 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7"} Mar 13 15:49:27 crc kubenswrapper[4907]: I0313 15:49:27.800040 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerStarted","Data":"5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797"} Mar 13 15:49:27 crc kubenswrapper[4907]: I0313 15:49:27.800606 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 15:49:27 crc kubenswrapper[4907]: I0313 15:49:27.829619 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.415617243 podStartE2EDuration="5.82959926s" podCreationTimestamp="2026-03-13 15:49:22 +0000 UTC" firstStartedPulling="2026-03-13 15:49:23.027813188 +0000 UTC m=+6261.927600877" lastFinishedPulling="2026-03-13 15:49:27.441795205 +0000 UTC m=+6266.341582894" observedRunningTime="2026-03-13 15:49:27.82117969 +0000 UTC m=+6266.720967379" watchObservedRunningTime="2026-03-13 15:49:27.82959926 +0000 UTC m=+6266.729386949" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.386563 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.388448 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.397612 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.417030 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.417475 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.519729 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.519901 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.520741 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.545523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"aodh-db-create-6qbgl\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.595684 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.597249 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.602375 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.611649 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.711760 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.723912 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.724007 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.832830 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.832968 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.840830 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.880169 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"aodh-cc97-account-create-update-fc5r7\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:33 crc kubenswrapper[4907]: I0313 15:49:33.920481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.303209 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:49:34 crc kubenswrapper[4907]: W0313 15:49:34.663319 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod442daa10_2973_4304_9e1c_039d35226686.slice/crio-42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d WatchSource:0}: Error finding container 42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d: Status 404 returned error can't find the container with id 42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.677778 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.880667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cc97-account-create-update-fc5r7" event={"ID":"442daa10-2973-4304-9e1c-039d35226686","Type":"ContainerStarted","Data":"42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d"} Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.882186 4907 generic.go:334] "Generic (PLEG): container finished" podID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerID="7eeac1862f5788609b87b66dfacb73f343af69811adf4dfc0c22454194586bf3" exitCode=0 Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.882256 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6qbgl" event={"ID":"689fb651-0113-48ba-b6e4-29c41c5fa21f","Type":"ContainerDied","Data":"7eeac1862f5788609b87b66dfacb73f343af69811adf4dfc0c22454194586bf3"} Mar 13 15:49:34 crc kubenswrapper[4907]: I0313 15:49:34.882352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6qbgl" event={"ID":"689fb651-0113-48ba-b6e4-29c41c5fa21f","Type":"ContainerStarted","Data":"36da2d4a680e719ad1fe919746dbf7987f40a93fe7fec96a6662addf2879d7df"} Mar 13 15:49:35 crc kubenswrapper[4907]: I0313 15:49:35.892135 4907 generic.go:334] "Generic (PLEG): container finished" podID="442daa10-2973-4304-9e1c-039d35226686" containerID="d38dedf8f9062f909f22038364138fce749f1c07518101c6bfdc3e1d5f365369" exitCode=0 Mar 13 15:49:35 crc kubenswrapper[4907]: I0313 15:49:35.892182 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cc97-account-create-update-fc5r7" event={"ID":"442daa10-2973-4304-9e1c-039d35226686","Type":"ContainerDied","Data":"d38dedf8f9062f909f22038364138fce749f1c07518101c6bfdc3e1d5f365369"} Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.309817 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.393610 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") pod \"689fb651-0113-48ba-b6e4-29c41c5fa21f\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.393728 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") pod \"689fb651-0113-48ba-b6e4-29c41c5fa21f\" (UID: \"689fb651-0113-48ba-b6e4-29c41c5fa21f\") " Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.394403 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "689fb651-0113-48ba-b6e4-29c41c5fa21f" (UID: "689fb651-0113-48ba-b6e4-29c41c5fa21f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.394599 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/689fb651-0113-48ba-b6e4-29c41c5fa21f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.399958 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z" (OuterVolumeSpecName: "kube-api-access-r7s7z") pod "689fb651-0113-48ba-b6e4-29c41c5fa21f" (UID: "689fb651-0113-48ba-b6e4-29c41c5fa21f"). InnerVolumeSpecName "kube-api-access-r7s7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.496973 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7s7z\" (UniqueName: \"kubernetes.io/projected/689fb651-0113-48ba-b6e4-29c41c5fa21f-kube-api-access-r7s7z\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.906007 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-6qbgl" event={"ID":"689fb651-0113-48ba-b6e4-29c41c5fa21f","Type":"ContainerDied","Data":"36da2d4a680e719ad1fe919746dbf7987f40a93fe7fec96a6662addf2879d7df"} Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.906069 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36da2d4a680e719ad1fe919746dbf7987f40a93fe7fec96a6662addf2879d7df" Mar 13 15:49:36 crc kubenswrapper[4907]: I0313 15:49:36.906131 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-6qbgl" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.328666 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.418213 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") pod \"442daa10-2973-4304-9e1c-039d35226686\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.418345 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") pod \"442daa10-2973-4304-9e1c-039d35226686\" (UID: \"442daa10-2973-4304-9e1c-039d35226686\") " Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.418845 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "442daa10-2973-4304-9e1c-039d35226686" (UID: "442daa10-2973-4304-9e1c-039d35226686"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.419490 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/442daa10-2973-4304-9e1c-039d35226686-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.443649 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw" (OuterVolumeSpecName: "kube-api-access-74mmw") pod "442daa10-2973-4304-9e1c-039d35226686" (UID: "442daa10-2973-4304-9e1c-039d35226686"). InnerVolumeSpecName "kube-api-access-74mmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.522603 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74mmw\" (UniqueName: \"kubernetes.io/projected/442daa10-2973-4304-9e1c-039d35226686-kube-api-access-74mmw\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.578752 4907 scope.go:117] "RemoveContainer" containerID="10110c9c9bf7e0c30409af7a457a19b276708bc01363735a9e6dff6935d07eb0" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.620170 4907 scope.go:117] "RemoveContainer" containerID="48b0748e550ebe6b33142e46ea370e9b43cfd8df0c8482529bda2ee099868df0" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.658649 4907 scope.go:117] "RemoveContainer" containerID="dcfa157fb4134a5c5ee1865cf6f3b6a035a94e4eeb4f7a317b410b36548c37aa" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.917392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-cc97-account-create-update-fc5r7" event={"ID":"442daa10-2973-4304-9e1c-039d35226686","Type":"ContainerDied","Data":"42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d"} Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.917699 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42a5129d078b177eff9fafc5af8936456bf22055b57fe17bc2a77ae226405e6d" Mar 13 15:49:37 crc kubenswrapper[4907]: I0313 15:49:37.917587 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-cc97-account-create-update-fc5r7" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.880442 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:49:38 crc kubenswrapper[4907]: E0313 15:49:38.880866 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerName="mariadb-database-create" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.880899 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerName="mariadb-database-create" Mar 13 15:49:38 crc kubenswrapper[4907]: E0313 15:49:38.880914 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442daa10-2973-4304-9e1c-039d35226686" containerName="mariadb-account-create-update" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.880920 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="442daa10-2973-4304-9e1c-039d35226686" containerName="mariadb-account-create-update" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.881293 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="442daa10-2973-4304-9e1c-039d35226686" containerName="mariadb-account-create-update" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.881310 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" containerName="mariadb-database-create" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.882039 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886124 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886300 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6c7x" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886329 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.886440 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.896080 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.951433 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.951642 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.951735 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:38 crc kubenswrapper[4907]: I0313 15:49:38.952024 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.053953 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.054071 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.054172 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.054224 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.060528 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.062343 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.066566 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.081992 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"aodh-db-sync-nwwc8\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.202558 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.677869 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:49:39 crc kubenswrapper[4907]: I0313 15:49:39.948968 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerStarted","Data":"5d0309bb66d24476037e34dfde5f9f67db1d48d62b2ca5025abb9b5b8f5d7cb3"} Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.110656 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerStarted","Data":"84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a"} Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.144519 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-nwwc8" podStartSLOduration=2.779150826 podStartE2EDuration="7.144496635s" podCreationTimestamp="2026-03-13 15:49:38 +0000 UTC" firstStartedPulling="2026-03-13 15:49:39.693870674 +0000 UTC m=+6278.593658353" lastFinishedPulling="2026-03-13 15:49:44.059216473 +0000 UTC m=+6282.959004162" observedRunningTime="2026-03-13 15:49:45.139740874 +0000 UTC m=+6284.039528573" watchObservedRunningTime="2026-03-13 15:49:45.144496635 +0000 UTC m=+6284.044284324" Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.735635 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-bwznz" podUID="e3a3b609-3625-4670-b669-d7fd07386be6" containerName="registry-server" probeResult="failure" output=< Mar 13 15:49:45 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:49:45 crc kubenswrapper[4907]: > Mar 13 15:49:45 crc kubenswrapper[4907]: I0313 15:49:45.735931 4907 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-bwznz" podUID="e3a3b609-3625-4670-b669-d7fd07386be6" containerName="registry-server" probeResult="failure" output=< Mar 13 15:49:45 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:49:45 crc kubenswrapper[4907]: > Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.041470 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.042832 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.042983 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.044005 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.044144 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2" gracePeriod=600 Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.143350 4907 generic.go:334] "Generic (PLEG): container finished" podID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerID="84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a" exitCode=0 Mar 13 15:49:48 crc kubenswrapper[4907]: I0313 15:49:48.143393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerDied","Data":"84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a"} Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.156388 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2" exitCode=0 Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.157127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2"} Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.157155 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792"} Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.157170 4907 scope.go:117] "RemoveContainer" containerID="68ffe92dd03078633ec89d7ef29da9bbdc4c5f04c956a1e8487e3f2e01c5d1c5" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.677359 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746329 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746679 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746778 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.746823 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") pod \"c5a77890-eb92-4d55-acf8-481f70c9277c\" (UID: \"c5a77890-eb92-4d55-acf8-481f70c9277c\") " Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.751496 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh" (OuterVolumeSpecName: "kube-api-access-rt9sh") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "kube-api-access-rt9sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.753153 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts" (OuterVolumeSpecName: "scripts") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.773787 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data" (OuterVolumeSpecName: "config-data") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.774969 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5a77890-eb92-4d55-acf8-481f70c9277c" (UID: "c5a77890-eb92-4d55-acf8-481f70c9277c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849273 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849303 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849313 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt9sh\" (UniqueName: \"kubernetes.io/projected/c5a77890-eb92-4d55-acf8-481f70c9277c-kube-api-access-rt9sh\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:49 crc kubenswrapper[4907]: I0313 15:49:49.849321 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5a77890-eb92-4d55-acf8-481f70c9277c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.171511 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-nwwc8" event={"ID":"c5a77890-eb92-4d55-acf8-481f70c9277c","Type":"ContainerDied","Data":"5d0309bb66d24476037e34dfde5f9f67db1d48d62b2ca5025abb9b5b8f5d7cb3"} Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.171834 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d0309bb66d24476037e34dfde5f9f67db1d48d62b2ca5025abb9b5b8f5d7cb3" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.171574 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-nwwc8" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.767564 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:49:50 crc kubenswrapper[4907]: E0313 15:49:50.768516 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerName="aodh-db-sync" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.768541 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerName="aodh-db-sync" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.768777 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" containerName="aodh-db-sync" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.770324 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.791032 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.867113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.867165 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.867209 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.969984 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.970027 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.970078 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.971249 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.971262 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:50 crc kubenswrapper[4907]: I0313 15:49:50.994655 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"certified-operators-jj9xr\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:51 crc kubenswrapper[4907]: I0313 15:49:51.096161 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:49:51 crc kubenswrapper[4907]: I0313 15:49:51.599927 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.199462 4907 generic.go:334] "Generic (PLEG): container finished" podID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" exitCode=0 Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.199523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999"} Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.199807 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerStarted","Data":"02bff9741d914eb2d1510a0ab7003a1cb31cbd80634618a50cf32118561743ca"} Mar 13 15:49:52 crc kubenswrapper[4907]: I0313 15:49:52.546148 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 15:49:53 crc kubenswrapper[4907]: I0313 15:49:53.212304 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerStarted","Data":"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05"} Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.040194 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.043446 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.046275 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-j6c7x" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.046950 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.047649 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.051780 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139345 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdszn\" (UniqueName: \"kubernetes.io/projected/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-kube-api-access-wdszn\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139417 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-scripts\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139486 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-config-data\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.139567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241374 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdszn\" (UniqueName: \"kubernetes.io/projected/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-kube-api-access-wdszn\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241438 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-scripts\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241505 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-config-data\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.241581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.253671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-scripts\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.253950 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.254698 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-config-data\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.268354 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdszn\" (UniqueName: \"kubernetes.io/projected/8ce53a96-2566-4d5c-b4cd-bbfd2516975a-kube-api-access-wdszn\") pod \"aodh-0\" (UID: \"8ce53a96-2566-4d5c-b4cd-bbfd2516975a\") " pod="openstack/aodh-0" Mar 13 15:49:54 crc kubenswrapper[4907]: I0313 15:49:54.364538 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 15:49:55 crc kubenswrapper[4907]: W0313 15:49:54.956069 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ce53a96_2566_4d5c_b4cd_bbfd2516975a.slice/crio-643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13 WatchSource:0}: Error finding container 643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13: Status 404 returned error can't find the container with id 643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13 Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:54.958421 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.048655 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.063112 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5r4tb"] Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.237747 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"643f6e0412f2540662588d14df1c54b0fff1ba0d761c6475ac478360aa911c13"} Mar 13 15:49:55 crc kubenswrapper[4907]: I0313 15:49:55.795307 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5aab76f-4bcc-47be-989d-7d9d6ecf371b" path="/var/lib/kubelet/pods/e5aab76f-4bcc-47be-989d-7d9d6ecf371b/volumes" Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.047679 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.064397 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.074394 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.086867 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-7djpm"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.095761 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.105287 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-62d0-account-create-update-vf8sr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.115159 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5a99-account-create-update-nkg4x"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.124522 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.133302 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-brbzj"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.141249 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-356d-account-create-update-kb8lr"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.247862 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"f540978f1e9f16494cf5a8dc9d6288ee493f15ca0f0d76fc59aee4c243f42647"} Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.338903 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.339308 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" containerID="cri-o://b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0" gracePeriod=30 Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.339758 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" containerID="cri-o://5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797" gracePeriod=30 Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.339834 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" containerID="cri-o://f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0" gracePeriod=30 Mar 13 15:49:56 crc kubenswrapper[4907]: I0313 15:49:56.340081 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" containerID="cri-o://3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7" gracePeriod=30 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.288406 4907 generic.go:334] "Generic (PLEG): container finished" podID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.288485 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305409 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305440 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7" exitCode=2 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305448 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305461 4907 generic.go:334] "Generic (PLEG): container finished" podID="eaec5907-c808-4ac2-b131-608b228885e6" containerID="b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0" exitCode=0 Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305482 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305510 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.305529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0"} Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.442192 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513515 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513600 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513737 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513801 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513868 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513970 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.513998 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") pod \"eaec5907-c808-4ac2-b131-608b228885e6\" (UID: \"eaec5907-c808-4ac2-b131-608b228885e6\") " Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.517475 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.517792 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.536765 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts" (OuterVolumeSpecName: "scripts") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.536876 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx" (OuterVolumeSpecName: "kube-api-access-x6gpx") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "kube-api-access-x6gpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.547468 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.616958 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.616996 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6gpx\" (UniqueName: \"kubernetes.io/projected/eaec5907-c808-4ac2-b131-608b228885e6-kube-api-access-x6gpx\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.617011 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.617021 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eaec5907-c808-4ac2-b131-608b228885e6-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.617031 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.648963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.651075 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data" (OuterVolumeSpecName: "config-data") pod "eaec5907-c808-4ac2-b131-608b228885e6" (UID: "eaec5907-c808-4ac2-b131-608b228885e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.718322 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.718350 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaec5907-c808-4ac2-b131-608b228885e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.799664 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18fa9304-50a2-45cf-b983-9d459f0dcb5f" path="/var/lib/kubelet/pods/18fa9304-50a2-45cf-b983-9d459f0dcb5f/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.809348 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40763271-36d9-4bd2-8ae8-82140648fcf4" path="/var/lib/kubelet/pods/40763271-36d9-4bd2-8ae8-82140648fcf4/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.810444 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94ba887c-f88a-48e7-b4bc-2f171d21fa47" path="/var/lib/kubelet/pods/94ba887c-f88a-48e7-b4bc-2f171d21fa47/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.811001 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b66093c1-e494-4975-b8d1-217844060204" path="/var/lib/kubelet/pods/b66093c1-e494-4975-b8d1-217844060204/volumes" Mar 13 15:49:57 crc kubenswrapper[4907]: I0313 15:49:57.812036 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c52fb5-c4ba-4c24-95af-ce75cd609126" path="/var/lib/kubelet/pods/c2c52fb5-c4ba-4c24-95af-ce75cd609126/volumes" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.317450 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerStarted","Data":"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37"} Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.322792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"646195359af961146eb2ff6e07795dd84d4ca2e44a087864e132385c5bbf36c8"} Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.326529 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eaec5907-c808-4ac2-b131-608b228885e6","Type":"ContainerDied","Data":"ae9b3358826e66742e51bbbc204d8b30a566e7e0b936fca1df7ae895ee7043e9"} Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.326582 4907 scope.go:117] "RemoveContainer" containerID="5a450b86aabb5b84d29c1e3252b0f33c4167256fbc37a7560f1e23aae3ba5797" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.326765 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.347777 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jj9xr" podStartSLOduration=2.58015895 podStartE2EDuration="8.34774993s" podCreationTimestamp="2026-03-13 15:49:50 +0000 UTC" firstStartedPulling="2026-03-13 15:49:52.20284506 +0000 UTC m=+6291.102632749" lastFinishedPulling="2026-03-13 15:49:57.97043604 +0000 UTC m=+6296.870223729" observedRunningTime="2026-03-13 15:49:58.333314825 +0000 UTC m=+6297.233102514" watchObservedRunningTime="2026-03-13 15:49:58.34774993 +0000 UTC m=+6297.247537619" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.367000 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.367963 4907 scope.go:117] "RemoveContainer" containerID="3b5b22ec27be798569c983dbf41141eab318a4ae784c5efd3348ea56c4f000f7" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.388673 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.401296 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402053 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402079 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402088 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402104 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402113 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: E0313 15:49:58.402137 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402146 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402430 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="sg-core" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402455 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-central-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402478 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="ceilometer-notification-agent" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.402491 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaec5907-c808-4ac2-b131-608b228885e6" containerName="proxy-httpd" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.405135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.406865 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.408092 4907 scope.go:117] "RemoveContainer" containerID="f8bccaf5228a1db9069e313bf3466a6f8cbd61fae45f55bda0a236da15bba7f0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.408263 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.408324 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.471221 4907 scope.go:117] "RemoveContainer" containerID="b3ab302efa65117dbfc68c5b768e1ee7a624666e299637f13e417c20da84fcf0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536211 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536261 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536284 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536378 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536415 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536449 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.536508 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638099 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638235 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638265 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638292 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638356 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.638971 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.639008 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.645975 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.647536 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.648986 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.654493 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.658523 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"ceilometer-0\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " pod="openstack/ceilometer-0" Mar 13 15:49:58 crc kubenswrapper[4907]: I0313 15:49:58.728916 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:49:59 crc kubenswrapper[4907]: I0313 15:49:59.360716 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:49:59 crc kubenswrapper[4907]: I0313 15:49:59.794002 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaec5907-c808-4ac2-b131-608b228885e6" path="/var/lib/kubelet/pods/eaec5907-c808-4ac2-b131-608b228885e6/volumes" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.134494 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.136372 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.139218 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.139490 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.144362 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.152696 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.172301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"auto-csr-approver-29556950-pqwgk\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.274612 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"auto-csr-approver-29556950-pqwgk\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.299062 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"auto-csr-approver-29556950-pqwgk\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.357205 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"179402b0df76b90aa1275176df179af7bc2ec062262c7c78fb13ace6f27c6838"} Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.363071 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9"} Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.363106 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"9ba05f967433eff9725a6dcc1ca7f6cab5e912da0e5ca7e888033a6d0b137522"} Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.474203 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:00 crc kubenswrapper[4907]: I0313 15:50:00.975074 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.096607 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.096992 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.161868 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.377543 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerStarted","Data":"df5d57d55555efa41307fa8944531c055731a32dc5db86d222d79f540686b70e"} Mar 13 15:50:01 crc kubenswrapper[4907]: I0313 15:50:01.379489 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.392573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerStarted","Data":"fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.398194 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"8ce53a96-2566-4d5c-b4cd-bbfd2516975a","Type":"ContainerStarted","Data":"75b45c6cebfb5cd8199369e1018195ba07a407a3aa02a55e64b34a452c02f794"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.400412 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092"} Mar 13 15:50:02 crc kubenswrapper[4907]: I0313 15:50:02.418656 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" podStartSLOduration=1.423557527 podStartE2EDuration="2.418640263s" podCreationTimestamp="2026-03-13 15:50:00 +0000 UTC" firstStartedPulling="2026-03-13 15:50:01.015037555 +0000 UTC m=+6299.914825244" lastFinishedPulling="2026-03-13 15:50:02.010120291 +0000 UTC m=+6300.909907980" observedRunningTime="2026-03-13 15:50:02.41742792 +0000 UTC m=+6301.317215609" watchObservedRunningTime="2026-03-13 15:50:02.418640263 +0000 UTC m=+6301.318427972" Mar 13 15:50:04 crc kubenswrapper[4907]: I0313 15:50:04.426790 4907 generic.go:334] "Generic (PLEG): container finished" podID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerID="fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0" exitCode=0 Mar 13 15:50:04 crc kubenswrapper[4907]: I0313 15:50:04.426890 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerDied","Data":"fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0"} Mar 13 15:50:04 crc kubenswrapper[4907]: I0313 15:50:04.444802 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.4641411570000002 podStartE2EDuration="10.44478282s" podCreationTimestamp="2026-03-13 15:49:54 +0000 UTC" firstStartedPulling="2026-03-13 15:49:54.958841099 +0000 UTC m=+6293.858628788" lastFinishedPulling="2026-03-13 15:50:01.939482762 +0000 UTC m=+6300.839270451" observedRunningTime="2026-03-13 15:50:02.443082941 +0000 UTC m=+6301.342870630" watchObservedRunningTime="2026-03-13 15:50:04.44478282 +0000 UTC m=+6303.344570509" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.035721 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.045297 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mwt47"] Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.438405 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerStarted","Data":"ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a"} Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.438903 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.464221 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.584673066 podStartE2EDuration="7.464197563s" podCreationTimestamp="2026-03-13 15:49:58 +0000 UTC" firstStartedPulling="2026-03-13 15:49:59.38505795 +0000 UTC m=+6298.284845639" lastFinishedPulling="2026-03-13 15:50:04.264582457 +0000 UTC m=+6303.164370136" observedRunningTime="2026-03-13 15:50:05.462577018 +0000 UTC m=+6304.362364717" watchObservedRunningTime="2026-03-13 15:50:05.464197563 +0000 UTC m=+6304.363985252" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.800787 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0336c3c0-cc18-4bb3-afe3-7e2a0232c889" path="/var/lib/kubelet/pods/0336c3c0-cc18-4bb3-afe3-7e2a0232c889/volumes" Mar 13 15:50:05 crc kubenswrapper[4907]: I0313 15:50:05.954159 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.009953 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") pod \"63d124ed-9ee2-437b-8ebd-b38a53e499af\" (UID: \"63d124ed-9ee2-437b-8ebd-b38a53e499af\") " Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.018262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k" (OuterVolumeSpecName: "kube-api-access-z6m5k") pod "63d124ed-9ee2-437b-8ebd-b38a53e499af" (UID: "63d124ed-9ee2-437b-8ebd-b38a53e499af"). InnerVolumeSpecName "kube-api-access-z6m5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.112427 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6m5k\" (UniqueName: \"kubernetes.io/projected/63d124ed-9ee2-437b-8ebd-b38a53e499af-kube-api-access-z6m5k\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.453141 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.453220 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556950-pqwgk" event={"ID":"63d124ed-9ee2-437b-8ebd-b38a53e499af","Type":"ContainerDied","Data":"df5d57d55555efa41307fa8944531c055731a32dc5db86d222d79f540686b70e"} Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.453775 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df5d57d55555efa41307fa8944531c055731a32dc5db86d222d79f540686b70e" Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.503310 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:50:06 crc kubenswrapper[4907]: I0313 15:50:06.512717 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556944-gcp6x"] Mar 13 15:50:07 crc kubenswrapper[4907]: I0313 15:50:07.797923 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04ca577b-ab8d-44a6-a598-1297bd5ab664" path="/var/lib/kubelet/pods/04ca577b-ab8d-44a6-a598-1297bd5ab664/volumes" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.977660 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 15:50:08 crc kubenswrapper[4907]: E0313 15:50:08.978296 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerName="oc" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.978316 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerName="oc" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.978564 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" containerName="oc" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.979521 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:08 crc kubenswrapper[4907]: I0313 15:50:08.987618 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.075526 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.075800 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.178370 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.178541 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.179356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.184615 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.185984 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.188717 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.200639 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"manila-db-create-mzct7\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.203059 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.284342 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.284482 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.301445 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.386212 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.386630 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.386990 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.406471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"manila-e749-account-create-update-gsxrn\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.568283 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:09 crc kubenswrapper[4907]: I0313 15:50:09.867464 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 15:50:10 crc kubenswrapper[4907]: W0313 15:50:10.082485 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0498e113_86c3_4718_a580_13bfdb277ff2.slice/crio-89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070 WatchSource:0}: Error finding container 89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070: Status 404 returned error can't find the container with id 89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070 Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.084506 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.507688 4907 generic.go:334] "Generic (PLEG): container finished" podID="b2a34745-a555-4029-a203-ee17762929e2" containerID="7fe743a125369b9e5af6f2cb7a4bc8d2add93450e84c6dadc626912538eac565" exitCode=0 Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.507752 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mzct7" event={"ID":"b2a34745-a555-4029-a203-ee17762929e2","Type":"ContainerDied","Data":"7fe743a125369b9e5af6f2cb7a4bc8d2add93450e84c6dadc626912538eac565"} Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.508101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mzct7" event={"ID":"b2a34745-a555-4029-a203-ee17762929e2","Type":"ContainerStarted","Data":"1b3b4efdf5f0221b588e062f2e104134c1f7d2650739a9addb8c08e1e1a79b65"} Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.509935 4907 generic.go:334] "Generic (PLEG): container finished" podID="0498e113-86c3-4718-a580-13bfdb277ff2" containerID="3da3591c8cb3c5a70194cd0e1a361e544712615fd8402b78ddd2f90160064c36" exitCode=0 Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.509986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e749-account-create-update-gsxrn" event={"ID":"0498e113-86c3-4718-a580-13bfdb277ff2","Type":"ContainerDied","Data":"3da3591c8cb3c5a70194cd0e1a361e544712615fd8402b78ddd2f90160064c36"} Mar 13 15:50:10 crc kubenswrapper[4907]: I0313 15:50:10.510013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e749-account-create-update-gsxrn" event={"ID":"0498e113-86c3-4718-a580-13bfdb277ff2","Type":"ContainerStarted","Data":"89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070"} Mar 13 15:50:11 crc kubenswrapper[4907]: I0313 15:50:11.153568 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:11 crc kubenswrapper[4907]: I0313 15:50:11.208708 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:50:11 crc kubenswrapper[4907]: I0313 15:50:11.518576 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jj9xr" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" containerID="cri-o://ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" gracePeriod=2 Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.092661 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.147814 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") pod \"0498e113-86c3-4718-a580-13bfdb277ff2\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.147952 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") pod \"0498e113-86c3-4718-a580-13bfdb277ff2\" (UID: \"0498e113-86c3-4718-a580-13bfdb277ff2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.149310 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0498e113-86c3-4718-a580-13bfdb277ff2" (UID: "0498e113-86c3-4718-a580-13bfdb277ff2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.155080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6" (OuterVolumeSpecName: "kube-api-access-xw8m6") pod "0498e113-86c3-4718-a580-13bfdb277ff2" (UID: "0498e113-86c3-4718-a580-13bfdb277ff2"). InnerVolumeSpecName "kube-api-access-xw8m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.244818 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.247373 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.251869 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0498e113-86c3-4718-a580-13bfdb277ff2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.252040 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw8m6\" (UniqueName: \"kubernetes.io/projected/0498e113-86c3-4718-a580-13bfdb277ff2-kube-api-access-xw8m6\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353145 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") pod \"b2a34745-a555-4029-a203-ee17762929e2\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353185 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") pod \"b2a34745-a555-4029-a203-ee17762929e2\" (UID: \"b2a34745-a555-4029-a203-ee17762929e2\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353230 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") pod \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353518 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") pod \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.353544 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") pod \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\" (UID: \"951bf99b-9f5e-4609-bff0-0c594f6a5b2e\") " Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.354393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2a34745-a555-4029-a203-ee17762929e2" (UID: "b2a34745-a555-4029-a203-ee17762929e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.354534 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities" (OuterVolumeSpecName: "utilities") pod "951bf99b-9f5e-4609-bff0-0c594f6a5b2e" (UID: "951bf99b-9f5e-4609-bff0-0c594f6a5b2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.357627 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2" (OuterVolumeSpecName: "kube-api-access-h9rf2") pod "951bf99b-9f5e-4609-bff0-0c594f6a5b2e" (UID: "951bf99b-9f5e-4609-bff0-0c594f6a5b2e"). InnerVolumeSpecName "kube-api-access-h9rf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.359130 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd" (OuterVolumeSpecName: "kube-api-access-2bvvd") pod "b2a34745-a555-4029-a203-ee17762929e2" (UID: "b2a34745-a555-4029-a203-ee17762929e2"). InnerVolumeSpecName "kube-api-access-2bvvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.405413 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "951bf99b-9f5e-4609-bff0-0c594f6a5b2e" (UID: "951bf99b-9f5e-4609-bff0-0c594f6a5b2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455895 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455926 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455937 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9rf2\" (UniqueName: \"kubernetes.io/projected/951bf99b-9f5e-4609-bff0-0c594f6a5b2e-kube-api-access-h9rf2\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455946 4907 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2a34745-a555-4029-a203-ee17762929e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.455955 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bvvd\" (UniqueName: \"kubernetes.io/projected/b2a34745-a555-4029-a203-ee17762929e2-kube-api-access-2bvvd\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.539183 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mzct7" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.539235 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mzct7" event={"ID":"b2a34745-a555-4029-a203-ee17762929e2","Type":"ContainerDied","Data":"1b3b4efdf5f0221b588e062f2e104134c1f7d2650739a9addb8c08e1e1a79b65"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.539630 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b3b4efdf5f0221b588e062f2e104134c1f7d2650739a9addb8c08e1e1a79b65" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.542567 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-e749-account-create-update-gsxrn" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.542607 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-e749-account-create-update-gsxrn" event={"ID":"0498e113-86c3-4718-a580-13bfdb277ff2","Type":"ContainerDied","Data":"89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.542659 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89c63ae209e5b88e5155699498019037c249d937a18a0b93309137427255d070" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545444 4907 generic.go:334] "Generic (PLEG): container finished" podID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" exitCode=0 Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545490 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545521 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jj9xr" event={"ID":"951bf99b-9f5e-4609-bff0-0c594f6a5b2e","Type":"ContainerDied","Data":"02bff9741d914eb2d1510a0ab7003a1cb31cbd80634618a50cf32118561743ca"} Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545540 4907 scope.go:117] "RemoveContainer" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.545536 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jj9xr" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.579155 4907 scope.go:117] "RemoveContainer" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.602553 4907 scope.go:117] "RemoveContainer" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.603693 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.626792 4907 scope.go:117] "RemoveContainer" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" Mar 13 15:50:12 crc kubenswrapper[4907]: E0313 15:50:12.627438 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37\": container with ID starting with ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37 not found: ID does not exist" containerID="ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.627501 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37"} err="failed to get container status \"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37\": rpc error: code = NotFound desc = could not find container \"ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37\": container with ID starting with ad22a18bbef271426a4077aab78c5dbcaeed67525c562d753181de83cefb8f37 not found: ID does not exist" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.627533 4907 scope.go:117] "RemoveContainer" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" Mar 13 15:50:12 crc kubenswrapper[4907]: E0313 15:50:12.628127 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05\": container with ID starting with d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05 not found: ID does not exist" containerID="d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.628185 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05"} err="failed to get container status \"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05\": rpc error: code = NotFound desc = could not find container \"d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05\": container with ID starting with d2ea7f99cadaa4f0d0bfca645932300ec0e331fafcb1e08d6511678f159fdd05 not found: ID does not exist" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.628217 4907 scope.go:117] "RemoveContainer" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" Mar 13 15:50:12 crc kubenswrapper[4907]: E0313 15:50:12.628548 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999\": container with ID starting with 240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999 not found: ID does not exist" containerID="240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.628580 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999"} err="failed to get container status \"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999\": rpc error: code = NotFound desc = could not find container \"240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999\": container with ID starting with 240a7e1dba69382ab9e9bd056e343991036dcabf16f06d6bad39768a4324c999 not found: ID does not exist" Mar 13 15:50:12 crc kubenswrapper[4907]: I0313 15:50:12.634644 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jj9xr"] Mar 13 15:50:13 crc kubenswrapper[4907]: I0313 15:50:13.795339 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" path="/var/lib/kubelet/pods/951bf99b-9f5e-4609-bff0-0c594f6a5b2e/volumes" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.525781 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526523 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526543 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526554 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" containerName="mariadb-account-create-update" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526564 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" containerName="mariadb-account-create-update" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526583 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a34745-a555-4029-a203-ee17762929e2" containerName="mariadb-database-create" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526594 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a34745-a555-4029-a203-ee17762929e2" containerName="mariadb-database-create" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526624 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-content" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526632 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-content" Mar 13 15:50:14 crc kubenswrapper[4907]: E0313 15:50:14.526668 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-utilities" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526678 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="extract-utilities" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526923 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a34745-a555-4029-a203-ee17762929e2" containerName="mariadb-database-create" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526962 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" containerName="mariadb-account-create-update" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.526976 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="951bf99b-9f5e-4609-bff0-0c594f6a5b2e" containerName="registry-server" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.527827 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.531414 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.531632 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-97kc6" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.550405 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.598828 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.598877 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.598978 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.599085 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.701463 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.702196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.702296 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.702325 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.708765 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.708891 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.709590 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.718746 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"manila-db-sync-nhx6q\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:14 crc kubenswrapper[4907]: I0313 15:50:14.862404 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:15 crc kubenswrapper[4907]: W0313 15:50:15.591276 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2df4eaaf_eb5c_4e04_b987_5aedb6561f7e.slice/crio-b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a WatchSource:0}: Error finding container b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a: Status 404 returned error can't find the container with id b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a Mar 13 15:50:15 crc kubenswrapper[4907]: I0313 15:50:15.591483 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 15:50:16 crc kubenswrapper[4907]: I0313 15:50:16.593479 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerStarted","Data":"b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a"} Mar 13 15:50:18 crc kubenswrapper[4907]: I0313 15:50:18.046968 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:50:18 crc kubenswrapper[4907]: I0313 15:50:18.061079 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-c2mhb"] Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.045925 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.062348 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lvbcr"] Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.793541 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89840ecf-ddb9-4a66-855b-0a1ae524bd6e" path="/var/lib/kubelet/pods/89840ecf-ddb9-4a66-855b-0a1ae524bd6e/volumes" Mar 13 15:50:19 crc kubenswrapper[4907]: I0313 15:50:19.877168 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e935140e-65a6-48ff-9dda-fdf57edb9a4d" path="/var/lib/kubelet/pods/e935140e-65a6-48ff-9dda-fdf57edb9a4d/volumes" Mar 13 15:50:21 crc kubenswrapper[4907]: I0313 15:50:21.646957 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerStarted","Data":"8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18"} Mar 13 15:50:21 crc kubenswrapper[4907]: I0313 15:50:21.671596 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-nhx6q" podStartSLOduration=2.719396925 podStartE2EDuration="7.671578957s" podCreationTimestamp="2026-03-13 15:50:14 +0000 UTC" firstStartedPulling="2026-03-13 15:50:15.593100602 +0000 UTC m=+6314.492888291" lastFinishedPulling="2026-03-13 15:50:20.545282634 +0000 UTC m=+6319.445070323" observedRunningTime="2026-03-13 15:50:21.665091169 +0000 UTC m=+6320.564878858" watchObservedRunningTime="2026-03-13 15:50:21.671578957 +0000 UTC m=+6320.571366646" Mar 13 15:50:23 crc kubenswrapper[4907]: I0313 15:50:23.664777 4907 generic.go:334] "Generic (PLEG): container finished" podID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerID="8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18" exitCode=0 Mar 13 15:50:23 crc kubenswrapper[4907]: I0313 15:50:23.664861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerDied","Data":"8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18"} Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.155648 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305116 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305277 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305447 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.305539 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") pod \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\" (UID: \"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e\") " Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.313189 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.314091 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj" (OuterVolumeSpecName: "kube-api-access-6hjjj") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "kube-api-access-6hjjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.315523 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data" (OuterVolumeSpecName: "config-data") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.337964 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" (UID: "2df4eaaf-eb5c-4e04-b987-5aedb6561f7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.407795 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hjjj\" (UniqueName: \"kubernetes.io/projected/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-kube-api-access-6hjjj\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.408248 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.408327 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.408402 4907 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e-job-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.702067 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-nhx6q" event={"ID":"2df4eaaf-eb5c-4e04-b987-5aedb6561f7e","Type":"ContainerDied","Data":"b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a"} Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.702125 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2c4c3a6559d1515e062025b3cd0e22f753423b868d9846b459513b583ca8b0a" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.702245 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-nhx6q" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.988516 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Mar 13 15:50:25 crc kubenswrapper[4907]: E0313 15:50:25.989125 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerName="manila-db-sync" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.989145 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerName="manila-db-sync" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.989374 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" containerName="manila-db-sync" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.990545 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.993852 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.994104 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.994176 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-97kc6" Mar 13 15:50:25 crc kubenswrapper[4907]: I0313 15:50:25.995559 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.001122 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.068574 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.070964 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.073535 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.101180 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.127935 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tkg6\" (UniqueName: \"kubernetes.io/projected/2e179ecf-0fc5-4675-89a6-287d3398dc2d-kube-api-access-7tkg6\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128192 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128231 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128279 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128306 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e179ecf-0fc5-4675-89a6-287d3398dc2d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.128374 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-scripts\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.183940 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.186172 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.198936 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235373 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbkz8\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-kube-api-access-tbkz8\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235532 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235570 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-scripts\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235849 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tkg6\" (UniqueName: \"kubernetes.io/projected/2e179ecf-0fc5-4675-89a6-287d3398dc2d-kube-api-access-7tkg6\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235940 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.235982 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236043 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236076 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e179ecf-0fc5-4675-89a6-287d3398dc2d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236136 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236187 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-scripts\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236227 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-ceph\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.236253 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.238660 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2e179ecf-0fc5-4675-89a6-287d3398dc2d-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.242521 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.243048 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.251603 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-config-data\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.258432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e179ecf-0fc5-4675-89a6-287d3398dc2d-scripts\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.258754 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tkg6\" (UniqueName: \"kubernetes.io/projected/2e179ecf-0fc5-4675-89a6-287d3398dc2d-kube-api-access-7tkg6\") pod \"manila-scheduler-0\" (UID: \"2e179ecf-0fc5-4675-89a6-287d3398dc2d\") " pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.322090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.337799 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338281 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-scripts\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338331 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338362 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338579 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338658 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-ceph\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338763 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbkz8\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-kube-api-access-tbkz8\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.338946 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.339118 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.341918 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/93911045-021f-489b-9817-f0a24066c906-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.348059 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-ceph\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.354072 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.356331 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.367167 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.369915 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.378085 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.379542 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.381227 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.381798 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93911045-021f-489b-9817-f0a24066c906-scripts\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.395152 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbkz8\" (UniqueName: \"kubernetes.io/projected/93911045-021f-489b-9817-f0a24066c906-kube-api-access-tbkz8\") pod \"manila-share-share1-0\" (UID: \"93911045-021f-489b-9817-f0a24066c906\") " pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440509 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440747 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440870 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.440996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.441110 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.442520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.442805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.443608 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.444224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.469030 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"dnsmasq-dns-7559d477c5-jggzg\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.507589 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543432 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2518181-dd4d-4d35-9377-061bbd91c811-etc-machine-id\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543468 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtp6p\" (UniqueName: \"kubernetes.io/projected/f2518181-dd4d-4d35-9377-061bbd91c811-kube-api-access-wtp6p\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-scripts\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543663 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data-custom\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.543694 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.544047 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2518181-dd4d-4d35-9377-061bbd91c811-logs\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645547 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2518181-dd4d-4d35-9377-061bbd91c811-logs\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645802 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2518181-dd4d-4d35-9377-061bbd91c811-etc-machine-id\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645830 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtp6p\" (UniqueName: \"kubernetes.io/projected/f2518181-dd4d-4d35-9377-061bbd91c811-kube-api-access-wtp6p\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645954 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-scripts\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.645992 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data-custom\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.646009 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.646896 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f2518181-dd4d-4d35-9377-061bbd91c811-etc-machine-id\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.647353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f2518181-dd4d-4d35-9377-061bbd91c811-logs\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.653174 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.655242 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-scripts\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.656793 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.659975 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f2518181-dd4d-4d35-9377-061bbd91c811-config-data-custom\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.668895 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtp6p\" (UniqueName: \"kubernetes.io/projected/f2518181-dd4d-4d35-9377-061bbd91c811-kube-api-access-wtp6p\") pod \"manila-api-0\" (UID: \"f2518181-dd4d-4d35-9377-061bbd91c811\") " pod="openstack/manila-api-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.695225 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Mar 13 15:50:26 crc kubenswrapper[4907]: I0313 15:50:26.832870 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.049904 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.076420 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e179ecf_0fc5_4675_89a6_287d3398dc2d.slice/crio-1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a WatchSource:0}: Error finding container 1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a: Status 404 returned error can't find the container with id 1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.238861 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.243194 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda01ed54c_beec_437b_a4fb_c4c62a407d76.slice/crio-9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1 WatchSource:0}: Error finding container 9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1: Status 404 returned error can't find the container with id 9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1 Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.488533 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93911045_021f_489b_9817_f0a24066c906.slice/crio-03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc WatchSource:0}: Error finding container 03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc: Status 404 returned error can't find the container with id 03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.495452 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.653068 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Mar 13 15:50:27 crc kubenswrapper[4907]: W0313 15:50:27.673324 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2518181_dd4d_4d35_9377_061bbd91c811.slice/crio-ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad WatchSource:0}: Error finding container ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad: Status 404 returned error can't find the container with id ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.747989 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"93911045-021f-489b-9817-f0a24066c906","Type":"ContainerStarted","Data":"03324e1e5a626682ca7e1985f811fb94aa7f345d5a58a62c8ac49866dcee12fc"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.751808 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f2518181-dd4d-4d35-9377-061bbd91c811","Type":"ContainerStarted","Data":"ab6f686c99c98e885820c86563b65060a971372055d87706c79e25a3c7cb42ad"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.755328 4907 generic.go:334] "Generic (PLEG): container finished" podID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" exitCode=0 Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.755389 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerDied","Data":"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.755410 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerStarted","Data":"9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1"} Mar 13 15:50:27 crc kubenswrapper[4907]: I0313 15:50:27.761334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2e179ecf-0fc5-4675-89a6-287d3398dc2d","Type":"ContainerStarted","Data":"1f1c2255067bd6fc9165910e84d3cb05d7d77bbad2543937a8f65a853ae24c2a"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.738727 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.808535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f2518181-dd4d-4d35-9377-061bbd91c811","Type":"ContainerStarted","Data":"5f1ed1ab0466cc73b3b8f1643b218bd51370eb20dc17ff9fab6a25c347de95be"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.808614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f2518181-dd4d-4d35-9377-061bbd91c811","Type":"ContainerStarted","Data":"5b2926df5bae903b244bc30e7e4d11c8e0b8953f321de39acc3a4dd0facc6050"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.808656 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.813221 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerStarted","Data":"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.836974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2e179ecf-0fc5-4675-89a6-287d3398dc2d","Type":"ContainerStarted","Data":"46b91298287e1c80e41badb65f95f6de2c660f1da51593026a7395f0692ceb8e"} Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.839051 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=2.839035584 podStartE2EDuration="2.839035584s" podCreationTimestamp="2026-03-13 15:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:50:28.836487434 +0000 UTC m=+6327.736275133" watchObservedRunningTime="2026-03-13 15:50:28.839035584 +0000 UTC m=+6327.738823273" Mar 13 15:50:28 crc kubenswrapper[4907]: I0313 15:50:28.876625 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" podStartSLOduration=2.8766038099999998 podStartE2EDuration="2.87660381s" podCreationTimestamp="2026-03-13 15:50:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:50:28.859561734 +0000 UTC m=+6327.759349423" watchObservedRunningTime="2026-03-13 15:50:28.87660381 +0000 UTC m=+6327.776391499" Mar 13 15:50:29 crc kubenswrapper[4907]: I0313 15:50:29.850701 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2e179ecf-0fc5-4675-89a6-287d3398dc2d","Type":"ContainerStarted","Data":"1a5fd5ea066c835cbc4a2587d84d4226edb7adf4e37526ac446862123a7cc55a"} Mar 13 15:50:29 crc kubenswrapper[4907]: I0313 15:50:29.851116 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:29 crc kubenswrapper[4907]: I0313 15:50:29.874474 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.8626676189999998 podStartE2EDuration="4.874456533s" podCreationTimestamp="2026-03-13 15:50:25 +0000 UTC" firstStartedPulling="2026-03-13 15:50:27.079849419 +0000 UTC m=+6325.979637108" lastFinishedPulling="2026-03-13 15:50:28.091638333 +0000 UTC m=+6326.991426022" observedRunningTime="2026-03-13 15:50:29.87178299 +0000 UTC m=+6328.771570689" watchObservedRunningTime="2026-03-13 15:50:29.874456533 +0000 UTC m=+6328.774244222" Mar 13 15:50:33 crc kubenswrapper[4907]: I0313 15:50:33.030756 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:50:33 crc kubenswrapper[4907]: I0313 15:50:33.039632 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-d9b97"] Mar 13 15:50:33 crc kubenswrapper[4907]: I0313 15:50:33.805975 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6e87a99-1744-4002-a08b-dabf8a97e48b" path="/var/lib/kubelet/pods/f6e87a99-1744-4002-a08b-dabf8a97e48b/volumes" Mar 13 15:50:34 crc kubenswrapper[4907]: I0313 15:50:34.905298 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"93911045-021f-489b-9817-f0a24066c906","Type":"ContainerStarted","Data":"99ecdb015c2877cd4051dfe0e106413112d1b3457e2928edbcb32905a4c80b35"} Mar 13 15:50:35 crc kubenswrapper[4907]: I0313 15:50:35.917713 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"93911045-021f-489b-9817-f0a24066c906","Type":"ContainerStarted","Data":"5a0ea1386dc8ac86951f18d7e6d31e118f0d651a291679d095103cc0e5e2b43e"} Mar 13 15:50:35 crc kubenswrapper[4907]: I0313 15:50:35.942688 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.9867187680000002 podStartE2EDuration="9.942455971s" podCreationTimestamp="2026-03-13 15:50:26 +0000 UTC" firstStartedPulling="2026-03-13 15:50:27.490488369 +0000 UTC m=+6326.390276058" lastFinishedPulling="2026-03-13 15:50:34.446225572 +0000 UTC m=+6333.346013261" observedRunningTime="2026-03-13 15:50:35.942446661 +0000 UTC m=+6334.842234360" watchObservedRunningTime="2026-03-13 15:50:35.942455971 +0000 UTC m=+6334.842243670" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.323751 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.510024 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.593784 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.594026 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" containerID="cri-o://04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" gracePeriod=10 Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.697973 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.938553 4907 generic.go:334] "Generic (PLEG): container finished" podID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" exitCode=0 Mar 13 15:50:36 crc kubenswrapper[4907]: I0313 15:50:36.938615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerDied","Data":"04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a"} Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.273121 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399523 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399732 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399794 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399849 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.399906 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") pod \"bb294da7-acf1-4984-a401-bbfc15ee97b7\" (UID: \"bb294da7-acf1-4984-a401-bbfc15ee97b7\") " Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.414124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt" (OuterVolumeSpecName: "kube-api-access-jrkzt") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "kube-api-access-jrkzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.464761 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.475615 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.479276 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config" (OuterVolumeSpecName: "config") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505021 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb294da7-acf1-4984-a401-bbfc15ee97b7" (UID: "bb294da7-acf1-4984-a401-bbfc15ee97b7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505910 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505940 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505953 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505965 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrkzt\" (UniqueName: \"kubernetes.io/projected/bb294da7-acf1-4984-a401-bbfc15ee97b7-kube-api-access-jrkzt\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.505977 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb294da7-acf1-4984-a401-bbfc15ee97b7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.797664 4907 scope.go:117] "RemoveContainer" containerID="3d3465a0228af3b611e5ca1d2d283d5739801f17bcfa78429f88dbe78da94add" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.831861 4907 scope.go:117] "RemoveContainer" containerID="219cbe1415b39a75ce8171bb61ada0538ae87eb0c8b470a088b03fdc9289f0b1" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.885532 4907 scope.go:117] "RemoveContainer" containerID="0948541af6cc99af21bd6ee88e49185bc604c42dbf31f8661316b429337a2c7f" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.928438 4907 scope.go:117] "RemoveContainer" containerID="7157d1dc97b867b0493ef78bf3ed0dea2af629c4ae1bace22c4fa35387dfbe4f" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.955738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.956954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d59c8ff75-hm7rf" event={"ID":"bb294da7-acf1-4984-a401-bbfc15ee97b7","Type":"ContainerDied","Data":"062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908"} Mar 13 15:50:37 crc kubenswrapper[4907]: I0313 15:50:37.957030 4907 scope.go:117] "RemoveContainer" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.000787 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.025897 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d59c8ff75-hm7rf"] Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.091201 4907 scope.go:117] "RemoveContainer" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: E0313 15:50:38.107055 4907 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_dnsmasq-dns_dnsmasq-dns-6d59c8ff75-hm7rf_openstack_bb294da7-acf1-4984-a401-bbfc15ee97b7_0 in pod sandbox 062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908 from index: no such id: '04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a'" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: E0313 15:50:38.107111 4907 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_dnsmasq-dns_dnsmasq-dns-6d59c8ff75-hm7rf_openstack_bb294da7-acf1-4984-a401-bbfc15ee97b7_0 in pod sandbox 062befe692770a942fb75e8e6fd540c55d77667654c81f8c1cedf1f5d48f5908 from index: no such id: '04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a'" containerID="04739d4af66a61e29f63286c048d55b12a764eb2ad7e3456b1fff0c9a19ff45a" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.107140 4907 scope.go:117] "RemoveContainer" containerID="e0422eb37278e23b8d6a6f631b14d7a520f0f17ab9fd9d3ee373391ac6276186" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.155696 4907 scope.go:117] "RemoveContainer" containerID="c3aa022d7c9308a16e2c36ea9708c7b13f760c092c5e791147b871724f8b2c96" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.184002 4907 scope.go:117] "RemoveContainer" containerID="fea279a9a11b2a38768139c4f75fb40e14428aafd4173f88229129062a569479" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.209082 4907 scope.go:117] "RemoveContainer" containerID="d3e4c590d97414f576974cec65543e931ac60a3bcf12387dbe9df0a1a39e2c28" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.237766 4907 scope.go:117] "RemoveContainer" containerID="9d15efc6444fb5cd991e1e807468f40b653afac37426c92c9f3e8218e37197cf" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.265388 4907 scope.go:117] "RemoveContainer" containerID="0bdd4c179fd6402b48e39e0498ccf620cd42daa3f95e10bf8dc41cdb2673e660" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.313343 4907 scope.go:117] "RemoveContainer" containerID="2e59f7eb348a0c98bd44ebc16e7f2b9007773ae7074558512509321e58dadcfc" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.372054 4907 scope.go:117] "RemoveContainer" containerID="04b92be87927c1c66e4f4d3a575acbb3afb714960c495c7fe45aa93f70f86d7e" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.433295 4907 scope.go:117] "RemoveContainer" containerID="c18542c3b44a6f6eb28a26dfdcfcff6b4ceb92d4f5c1a2a581f9880ab136aa90" Mar 13 15:50:38 crc kubenswrapper[4907]: I0313 15:50:38.488249 4907 scope.go:117] "RemoveContainer" containerID="f18dc1660972c456737f08b933ea252844a2bf45813b5a1dae54e428ebc166a5" Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.273619 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274229 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" containerID="cri-o://327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274363 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" containerID="cri-o://ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274410 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" containerID="cri-o://d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.274443 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" containerID="cri-o://34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d" gracePeriod=30 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.798413 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" path="/var/lib/kubelet/pods/bb294da7-acf1-4984-a401-bbfc15ee97b7/volumes" Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994437 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a" exitCode=0 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994471 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092" exitCode=2 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994480 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d" exitCode=0 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994488 4907 generic.go:334] "Generic (PLEG): container finished" podID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerID="327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9" exitCode=0 Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994508 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a"} Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092"} Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994545 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d"} Mar 13 15:50:39 crc kubenswrapper[4907]: I0313 15:50:39.994554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9"} Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.242349 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365048 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365149 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365237 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365272 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365303 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365326 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365396 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365655 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.365902 4907 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.366297 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.371542 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w" (OuterVolumeSpecName: "kube-api-access-ttn2w") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "kube-api-access-ttn2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.374552 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts" (OuterVolumeSpecName: "scripts") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.417235 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.466433 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.467367 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") pod \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\" (UID: \"14aa8b3a-f849-42c4-9248-2329fd28c6c3\") " Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.467988 4907 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468012 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttn2w\" (UniqueName: \"kubernetes.io/projected/14aa8b3a-f849-42c4-9248-2329fd28c6c3-kube-api-access-ttn2w\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468025 4907 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14aa8b3a-f849-42c4-9248-2329fd28c6c3-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468037 4907 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: W0313 15:50:40.468134 4907 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/14aa8b3a-f849-42c4-9248-2329fd28c6c3/volumes/kubernetes.io~secret/combined-ca-bundle Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.468149 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.475688 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data" (OuterVolumeSpecName: "config-data") pod "14aa8b3a-f849-42c4-9248-2329fd28c6c3" (UID: "14aa8b3a-f849-42c4-9248-2329fd28c6c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.570158 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:40 crc kubenswrapper[4907]: I0313 15:50:40.570205 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14aa8b3a-f849-42c4-9248-2329fd28c6c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.007057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14aa8b3a-f849-42c4-9248-2329fd28c6c3","Type":"ContainerDied","Data":"9ba05f967433eff9725a6dcc1ca7f6cab5e912da0e5ca7e888033a6d0b137522"} Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.007377 4907 scope.go:117] "RemoveContainer" containerID="ea745f91341a94f03f73269b8a99b71eff9cf760adb53889a4160dbc700df09a" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.007241 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.035140 4907 scope.go:117] "RemoveContainer" containerID="d5a88d5e47093bf38e94a5d416f23f1547843a36e587d1615bb482fff4d0a092" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.054051 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.065001 4907 scope.go:117] "RemoveContainer" containerID="34cd3ac68b3f061e9ab0b1f70d520ca0eebcf6f1d8afd0348c0574488177a64d" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.065071 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.083904 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084336 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084353 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084414 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084427 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084438 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084448 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084464 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084470 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084488 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084494 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" Mar 13 15:50:41 crc kubenswrapper[4907]: E0313 15:50:41.084525 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="init" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084532 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="init" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084758 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-notification-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084771 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="proxy-httpd" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084788 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="sg-core" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084801 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb294da7-acf1-4984-a401-bbfc15ee97b7" containerName="dnsmasq-dns" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.084810 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" containerName="ceilometer-central-agent" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.088001 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.090097 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.092850 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.097424 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.103642 4907 scope.go:117] "RemoveContainer" containerID="327a65f960da4ed89404c82b3f62aa43c7df17d2c7fdfd45430eda4484d631d9" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185511 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185559 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185660 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-scripts\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185683 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-log-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185699 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-config-data\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185789 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-run-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.185828 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svq8j\" (UniqueName: \"kubernetes.io/projected/59fb2cd9-5414-4356-9438-d33f90ad084c-kube-api-access-svq8j\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-run-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287667 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svq8j\" (UniqueName: \"kubernetes.io/projected/59fb2cd9-5414-4356-9438-d33f90ad084c-kube-api-access-svq8j\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287824 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287921 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-scripts\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287951 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-log-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.287973 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-config-data\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.288868 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-run-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.289309 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59fb2cd9-5414-4356-9438-d33f90ad084c-log-httpd\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.293861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.293977 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.294384 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-scripts\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.298077 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59fb2cd9-5414-4356-9438-d33f90ad084c-config-data\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.307421 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svq8j\" (UniqueName: \"kubernetes.io/projected/59fb2cd9-5414-4356-9438-d33f90ad084c-kube-api-access-svq8j\") pod \"ceilometer-0\" (UID: \"59fb2cd9-5414-4356-9438-d33f90ad084c\") " pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.406470 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.794489 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14aa8b3a-f849-42c4-9248-2329fd28c6c3" path="/var/lib/kubelet/pods/14aa8b3a-f849-42c4-9248-2329fd28c6c3/volumes" Mar 13 15:50:41 crc kubenswrapper[4907]: I0313 15:50:41.981682 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 15:50:42 crc kubenswrapper[4907]: I0313 15:50:42.017578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"0df3568657a2601e9a878c56a11954c34f6ef4dcfe1b9a5d4f12772552e8c314"} Mar 13 15:50:43 crc kubenswrapper[4907]: I0313 15:50:43.028190 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"7eea587aa87adf598d468a3ecc98d2f4ada4de324f142b4337f94eaa09b7c6ad"} Mar 13 15:50:44 crc kubenswrapper[4907]: I0313 15:50:44.049831 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"71de4d49169f6bb2dd395e4b26e3d015c4d648d12080086a842e391f8eef4e66"} Mar 13 15:50:45 crc kubenswrapper[4907]: I0313 15:50:45.062667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"351dde2349b18437c873dc5ab10acd990453b7071ad53b6cbdc2cd3f577a83c6"} Mar 13 15:50:47 crc kubenswrapper[4907]: I0313 15:50:47.123568 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59fb2cd9-5414-4356-9438-d33f90ad084c","Type":"ContainerStarted","Data":"6df1468393b247353225c6e4ccd83f9a2b2fb80610a9be665a4cea857f2ce364"} Mar 13 15:50:47 crc kubenswrapper[4907]: I0313 15:50:47.125209 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 15:50:47 crc kubenswrapper[4907]: I0313 15:50:47.155815 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.097447519 podStartE2EDuration="6.155794699s" podCreationTimestamp="2026-03-13 15:50:41 +0000 UTC" firstStartedPulling="2026-03-13 15:50:41.973807868 +0000 UTC m=+6340.873595557" lastFinishedPulling="2026-03-13 15:50:46.032155048 +0000 UTC m=+6344.931942737" observedRunningTime="2026-03-13 15:50:47.152827778 +0000 UTC m=+6346.052615467" watchObservedRunningTime="2026-03-13 15:50:47.155794699 +0000 UTC m=+6346.055582398" Mar 13 15:50:48 crc kubenswrapper[4907]: I0313 15:50:48.049269 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Mar 13 15:50:48 crc kubenswrapper[4907]: I0313 15:50:48.597577 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Mar 13 15:50:48 crc kubenswrapper[4907]: I0313 15:50:48.723556 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Mar 13 15:51:11 crc kubenswrapper[4907]: I0313 15:51:11.414322 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 15:51:16 crc kubenswrapper[4907]: I0313 15:51:16.055611 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:51:16 crc kubenswrapper[4907]: I0313 15:51:16.069123 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-9290-account-create-update-qws42"] Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.029814 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.043383 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-pvpwm"] Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.795569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ab8c7fa-9565-4b6c-b8f8-033cfdf444db" path="/var/lib/kubelet/pods/0ab8c7fa-9565-4b6c-b8f8-033cfdf444db/volumes" Mar 13 15:51:17 crc kubenswrapper[4907]: I0313 15:51:17.796846 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="499bc96f-8fa9-4900-bd92-2a507215791f" path="/var/lib/kubelet/pods/499bc96f-8fa9-4900-bd92-2a507215791f/volumes" Mar 13 15:51:24 crc kubenswrapper[4907]: I0313 15:51:24.046265 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:51:24 crc kubenswrapper[4907]: I0313 15:51:24.063745 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-hv2mf"] Mar 13 15:51:25 crc kubenswrapper[4907]: I0313 15:51:25.793178 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef105bad-6763-4234-b52b-6d2820d48b02" path="/var/lib/kubelet/pods/ef105bad-6763-4234-b52b-6d2820d48b02/volumes" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.302145 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.330610 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.330756 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.337734 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431108 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431633 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.431974 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.432054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.432159 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.534983 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535162 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535239 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535335 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535394 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.535602 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536340 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536455 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.536741 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.537094 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.569430 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"dnsmasq-dns-5ff7fb8bcc-45q4w\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:32 crc kubenswrapper[4907]: I0313 15:51:32.649666 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.121247 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.564227 4907 generic.go:334] "Generic (PLEG): container finished" podID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerID="bb89c1e3b4d030f795279c7977f6bc141b8f573b1af14ac2704526b6da21d04c" exitCode=0 Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.564293 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerDied","Data":"bb89c1e3b4d030f795279c7977f6bc141b8f573b1af14ac2704526b6da21d04c"} Mar 13 15:51:33 crc kubenswrapper[4907]: I0313 15:51:33.564351 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerStarted","Data":"5b3c0f84402a8ec16bbf16451e1bc1400975871789f821c116960d27711f8335"} Mar 13 15:51:34 crc kubenswrapper[4907]: I0313 15:51:34.577247 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerStarted","Data":"997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a"} Mar 13 15:51:34 crc kubenswrapper[4907]: I0313 15:51:34.577810 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:34 crc kubenswrapper[4907]: I0313 15:51:34.594589 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" podStartSLOduration=2.594569538 podStartE2EDuration="2.594569538s" podCreationTimestamp="2026-03-13 15:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:51:34.592232055 +0000 UTC m=+6393.492019764" watchObservedRunningTime="2026-03-13 15:51:34.594569538 +0000 UTC m=+6393.494357227" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.840676 4907 scope.go:117] "RemoveContainer" containerID="5140f9e51d2f4e22927da88ea13b8b76a7a069356e41a43262ca41c166a3779c" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.875557 4907 scope.go:117] "RemoveContainer" containerID="37e65533c5fc864e730a3583c9b7118a1665874961e44f735e26e376e6b5f39d" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.929815 4907 scope.go:117] "RemoveContainer" containerID="0559fde0f7d2daa5d36de2042211ef456904ceea0804232b86a43826c351c91c" Mar 13 15:51:38 crc kubenswrapper[4907]: I0313 15:51:38.971598 4907 scope.go:117] "RemoveContainer" containerID="cfb2b66bd790bad49fda00e0dee115de187df8f601026658c19c9100dc1ee877" Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.651101 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.708042 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.708290 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" containerID="cri-o://f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" gracePeriod=10 Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.901215 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8644fdb9df-gm62m"] Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.903073 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:42 crc kubenswrapper[4907]: I0313 15:51:42.915896 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8644fdb9df-gm62m"] Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069813 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9g6\" (UniqueName: \"kubernetes.io/projected/44fcf282-3688-4a6f-b390-6b321e89b158-kube-api-access-8v9g6\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069908 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-nb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069940 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-openstack-cell1\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.069962 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-sb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.070004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-dns-svc\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.070053 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-config\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172284 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9g6\" (UniqueName: \"kubernetes.io/projected/44fcf282-3688-4a6f-b390-6b321e89b158-kube-api-access-8v9g6\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172665 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-nb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-openstack-cell1\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172733 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-sb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172782 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-dns-svc\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.172844 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-config\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.174004 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-config\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.175974 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-sb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.176696 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-openstack-cell1\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.177491 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-dns-svc\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.179942 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44fcf282-3688-4a6f-b390-6b321e89b158-ovsdbserver-nb\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.193759 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9g6\" (UniqueName: \"kubernetes.io/projected/44fcf282-3688-4a6f-b390-6b321e89b158-kube-api-access-8v9g6\") pod \"dnsmasq-dns-8644fdb9df-gm62m\" (UID: \"44fcf282-3688-4a6f-b390-6b321e89b158\") " pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.279126 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.402561 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.481708 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.481777 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.482526 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.482624 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.483095 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") pod \"a01ed54c-beec-437b-a4fb-c4c62a407d76\" (UID: \"a01ed54c-beec-437b-a4fb-c4c62a407d76\") " Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.664939 4907 generic.go:334] "Generic (PLEG): container finished" podID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" exitCode=0 Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.664983 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.664986 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerDied","Data":"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f"} Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.665017 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7559d477c5-jggzg" event={"ID":"a01ed54c-beec-437b-a4fb-c4c62a407d76","Type":"ContainerDied","Data":"9777020954fe435146b223f1e22d2da8657fcd1b5305e36c2273a7c3f9c349b1"} Mar 13 15:51:43 crc kubenswrapper[4907]: I0313 15:51:43.665040 4907 scope.go:117] "RemoveContainer" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.093608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5" (OuterVolumeSpecName: "kube-api-access-bnxr5") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "kube-api-access-bnxr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.098764 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnxr5\" (UniqueName: \"kubernetes.io/projected/a01ed54c-beec-437b-a4fb-c4c62a407d76-kube-api-access-bnxr5\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.147296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config" (OuterVolumeSpecName: "config") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.149419 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.165280 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.176674 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a01ed54c-beec-437b-a4fb-c4c62a407d76" (UID: "a01ed54c-beec-437b-a4fb-c4c62a407d76"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200653 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200698 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200708 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.200719 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a01ed54c-beec-437b-a4fb-c4c62a407d76-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.305541 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8644fdb9df-gm62m"] Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.334644 4907 scope.go:117] "RemoveContainer" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.344562 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.356554 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7559d477c5-jggzg"] Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.374079 4907 scope.go:117] "RemoveContainer" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" Mar 13 15:51:44 crc kubenswrapper[4907]: E0313 15:51:44.374590 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f\": container with ID starting with f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f not found: ID does not exist" containerID="f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.374626 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f"} err="failed to get container status \"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f\": rpc error: code = NotFound desc = could not find container \"f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f\": container with ID starting with f427c6c6d4cea6920aec5e3d13f7c210e91731dd6a8240743203f4fd1462ee2f not found: ID does not exist" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.374646 4907 scope.go:117] "RemoveContainer" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" Mar 13 15:51:44 crc kubenswrapper[4907]: E0313 15:51:44.375002 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636\": container with ID starting with 6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636 not found: ID does not exist" containerID="6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.375026 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636"} err="failed to get container status \"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636\": rpc error: code = NotFound desc = could not find container \"6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636\": container with ID starting with 6a9689f3bb28f496b72b2c47ce88ac358b381c9e951286ab8db719b6cd0bd636 not found: ID does not exist" Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.676522 4907 generic.go:334] "Generic (PLEG): container finished" podID="44fcf282-3688-4a6f-b390-6b321e89b158" containerID="287d6d7b61fca2739c775824ecadcb069c5c0686f2843b53995e1578a705d92d" exitCode=0 Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.676560 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" event={"ID":"44fcf282-3688-4a6f-b390-6b321e89b158","Type":"ContainerDied","Data":"287d6d7b61fca2739c775824ecadcb069c5c0686f2843b53995e1578a705d92d"} Mar 13 15:51:44 crc kubenswrapper[4907]: I0313 15:51:44.676583 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" event={"ID":"44fcf282-3688-4a6f-b390-6b321e89b158","Type":"ContainerStarted","Data":"21b3b48967ffccd6fd70cace9989429623d571f5726ff15b165aad899cca0333"} Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.687999 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" event={"ID":"44fcf282-3688-4a6f-b390-6b321e89b158","Type":"ContainerStarted","Data":"115cd9adc28b0ba5e541b1c130a3291150179c7d1f979fd6c223d918a88300b5"} Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.689094 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.711051 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" podStartSLOduration=3.711034649 podStartE2EDuration="3.711034649s" podCreationTimestamp="2026-03-13 15:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 15:51:45.705509028 +0000 UTC m=+6404.605296717" watchObservedRunningTime="2026-03-13 15:51:45.711034649 +0000 UTC m=+6404.610822338" Mar 13 15:51:45 crc kubenswrapper[4907]: I0313 15:51:45.793554 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" path="/var/lib/kubelet/pods/a01ed54c-beec-437b-a4fb-c4c62a407d76/volumes" Mar 13 15:51:48 crc kubenswrapper[4907]: I0313 15:51:48.041623 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:51:48 crc kubenswrapper[4907]: I0313 15:51:48.042188 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.280045 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8644fdb9df-gm62m" Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.348680 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.348959 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" containerID="cri-o://997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a" gracePeriod=10 Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.794592 4907 generic.go:334] "Generic (PLEG): container finished" podID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerID="997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a" exitCode=0 Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.797819 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerDied","Data":"997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a"} Mar 13 15:51:53 crc kubenswrapper[4907]: I0313 15:51:53.902373 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.026979 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027032 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027175 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027300 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027340 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.027476 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") pod \"68d11f2a-37b6-4a6c-b877-ae32e177761d\" (UID: \"68d11f2a-37b6-4a6c-b877-ae32e177761d\") " Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.032981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s" (OuterVolumeSpecName: "kube-api-access-vww7s") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "kube-api-access-vww7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.085467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.093105 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.094537 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config" (OuterVolumeSpecName: "config") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.095456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.121357 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "68d11f2a-37b6-4a6c-b877-ae32e177761d" (UID: "68d11f2a-37b6-4a6c-b877-ae32e177761d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130222 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130271 4907 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-config\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130283 4907 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130294 4907 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130309 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vww7s\" (UniqueName: \"kubernetes.io/projected/68d11f2a-37b6-4a6c-b877-ae32e177761d-kube-api-access-vww7s\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.130324 4907 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68d11f2a-37b6-4a6c-b877-ae32e177761d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.805364 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" event={"ID":"68d11f2a-37b6-4a6c-b877-ae32e177761d","Type":"ContainerDied","Data":"5b3c0f84402a8ec16bbf16451e1bc1400975871789f821c116960d27711f8335"} Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.805416 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff7fb8bcc-45q4w" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.805677 4907 scope.go:117] "RemoveContainer" containerID="997f7ae74c7025a1a37f82e6b1e89fdb84b9a3fd4054caac0d2b4e2d822d549a" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.840548 4907 scope.go:117] "RemoveContainer" containerID="bb89c1e3b4d030f795279c7977f6bc141b8f573b1af14ac2704526b6da21d04c" Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.850320 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:54 crc kubenswrapper[4907]: I0313 15:51:54.872028 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ff7fb8bcc-45q4w"] Mar 13 15:51:55 crc kubenswrapper[4907]: I0313 15:51:55.793598 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" path="/var/lib/kubelet/pods/68d11f2a-37b6-4a6c-b877-ae32e177761d/volumes" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.142346 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143035 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143048 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143067 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143073 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143097 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143103 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="init" Mar 13 15:52:00 crc kubenswrapper[4907]: E0313 15:52:00.143116 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143122 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143296 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01ed54c-beec-437b-a4fb-c4c62a407d76" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143306 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d11f2a-37b6-4a6c-b877-ae32e177761d" containerName="dnsmasq-dns" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.143997 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.154069 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.199503 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.199706 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.199859 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.302055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"auto-csr-approver-29556952-b69cd\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.404801 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"auto-csr-approver-29556952-b69cd\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.427025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"auto-csr-approver-29556952-b69cd\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.519082 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:00 crc kubenswrapper[4907]: I0313 15:52:00.972482 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:52:01 crc kubenswrapper[4907]: I0313 15:52:01.876365 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556952-b69cd" event={"ID":"d2915500-b9a0-45a4-960e-5325f53b445c","Type":"ContainerStarted","Data":"427c240f7d555a5e5b1cbc79551c7beea9222cc99ef5a82bd5d85f606af5e2b4"} Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.261448 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw"] Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.263271 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.265219 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.267390 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.267659 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.267809 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.285241 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw"] Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.388578 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389099 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389212 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389469 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.389564 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491247 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491285 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491410 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.491444 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.497607 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.498510 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.499587 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.501192 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.513497 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.591749 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.903770 4907 generic.go:334] "Generic (PLEG): container finished" podID="d2915500-b9a0-45a4-960e-5325f53b445c" containerID="ae08e1f5613f0f2169631235f70419712208d7aa173e1b4d3da70be8496080b9" exitCode=0 Mar 13 15:52:04 crc kubenswrapper[4907]: I0313 15:52:04.904131 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556952-b69cd" event={"ID":"d2915500-b9a0-45a4-960e-5325f53b445c","Type":"ContainerDied","Data":"ae08e1f5613f0f2169631235f70419712208d7aa173e1b4d3da70be8496080b9"} Mar 13 15:52:05 crc kubenswrapper[4907]: I0313 15:52:05.194526 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw"] Mar 13 15:52:05 crc kubenswrapper[4907]: W0313 15:52:05.201355 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod044c1d1f_5bd4_4359_9262_e08daf71a4ae.slice/crio-40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8 WatchSource:0}: Error finding container 40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8: Status 404 returned error can't find the container with id 40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8 Mar 13 15:52:05 crc kubenswrapper[4907]: I0313 15:52:05.922794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerStarted","Data":"40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8"} Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.300114 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.437894 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") pod \"d2915500-b9a0-45a4-960e-5325f53b445c\" (UID: \"d2915500-b9a0-45a4-960e-5325f53b445c\") " Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.446621 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7" (OuterVolumeSpecName: "kube-api-access-q6vc7") pod "d2915500-b9a0-45a4-960e-5325f53b445c" (UID: "d2915500-b9a0-45a4-960e-5325f53b445c"). InnerVolumeSpecName "kube-api-access-q6vc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.540954 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6vc7\" (UniqueName: \"kubernetes.io/projected/d2915500-b9a0-45a4-960e-5325f53b445c-kube-api-access-q6vc7\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.956690 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556952-b69cd" event={"ID":"d2915500-b9a0-45a4-960e-5325f53b445c","Type":"ContainerDied","Data":"427c240f7d555a5e5b1cbc79551c7beea9222cc99ef5a82bd5d85f606af5e2b4"} Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.957010 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="427c240f7d555a5e5b1cbc79551c7beea9222cc99ef5a82bd5d85f606af5e2b4" Mar 13 15:52:06 crc kubenswrapper[4907]: I0313 15:52:06.956917 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556952-b69cd" Mar 13 15:52:07 crc kubenswrapper[4907]: I0313 15:52:07.375856 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:52:07 crc kubenswrapper[4907]: I0313 15:52:07.386359 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556946-78rqg"] Mar 13 15:52:07 crc kubenswrapper[4907]: I0313 15:52:07.798076 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f0d1531-1d25-4331-a690-1fcbcb504071" path="/var/lib/kubelet/pods/2f0d1531-1d25-4331-a690-1fcbcb504071/volumes" Mar 13 15:52:15 crc kubenswrapper[4907]: I0313 15:52:15.035736 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerStarted","Data":"a98ef489a033787cdd255f40580b6674e782d47c8d6a6991c57a495e6e6e8aad"} Mar 13 15:52:15 crc kubenswrapper[4907]: I0313 15:52:15.067069 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" podStartSLOduration=2.208048563 podStartE2EDuration="11.067034696s" podCreationTimestamp="2026-03-13 15:52:04 +0000 UTC" firstStartedPulling="2026-03-13 15:52:05.209144362 +0000 UTC m=+6424.108932051" lastFinishedPulling="2026-03-13 15:52:14.068130495 +0000 UTC m=+6432.967918184" observedRunningTime="2026-03-13 15:52:15.062509323 +0000 UTC m=+6433.962297032" watchObservedRunningTime="2026-03-13 15:52:15.067034696 +0000 UTC m=+6433.966822375" Mar 13 15:52:18 crc kubenswrapper[4907]: I0313 15:52:18.041380 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:52:18 crc kubenswrapper[4907]: I0313 15:52:18.041930 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:52:28 crc kubenswrapper[4907]: I0313 15:52:28.154028 4907 generic.go:334] "Generic (PLEG): container finished" podID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerID="a98ef489a033787cdd255f40580b6674e782d47c8d6a6991c57a495e6e6e8aad" exitCode=0 Mar 13 15:52:28 crc kubenswrapper[4907]: I0313 15:52:28.154115 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerDied","Data":"a98ef489a033787cdd255f40580b6674e782d47c8d6a6991c57a495e6e6e8aad"} Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.658013 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781107 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781196 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781691 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781843 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.781917 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") pod \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\" (UID: \"044c1d1f-5bd4-4359-9262-e08daf71a4ae\") " Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.787442 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28" (OuterVolumeSpecName: "kube-api-access-khw28") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "kube-api-access-khw28". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.787737 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.789207 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph" (OuterVolumeSpecName: "ceph") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.816733 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.841080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory" (OuterVolumeSpecName: "inventory") pod "044c1d1f-5bd4-4359-9262-e08daf71a4ae" (UID: "044c1d1f-5bd4-4359-9262-e08daf71a4ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886315 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886370 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khw28\" (UniqueName: \"kubernetes.io/projected/044c1d1f-5bd4-4359-9262-e08daf71a4ae-kube-api-access-khw28\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886386 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886399 4907 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:29 crc kubenswrapper[4907]: I0313 15:52:29.886415 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/044c1d1f-5bd4-4359-9262-e08daf71a4ae-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 15:52:30 crc kubenswrapper[4907]: I0313 15:52:30.177665 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" event={"ID":"044c1d1f-5bd4-4359-9262-e08daf71a4ae","Type":"ContainerDied","Data":"40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8"} Mar 13 15:52:30 crc kubenswrapper[4907]: I0313 15:52:30.178108 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40e9313e8c34a1e8fd1dc1058b1ee1e7934bed65552e6326335a68a5db3bf4f8" Mar 13 15:52:30 crc kubenswrapper[4907]: I0313 15:52:30.178057 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.992524 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz"] Mar 13 15:52:36 crc kubenswrapper[4907]: E0313 15:52:36.994387 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.994492 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Mar 13 15:52:36 crc kubenswrapper[4907]: E0313 15:52:36.994563 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" containerName="oc" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.994619 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" containerName="oc" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.995061 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" containerName="oc" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.995146 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="044c1d1f-5bd4-4359-9262-e08daf71a4ae" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.996072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:36 crc kubenswrapper[4907]: I0313 15:52:36.999038 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.004480 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.004707 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.004722 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.005831 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz"] Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.147770 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.148758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.149060 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.149144 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.149217 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251215 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251336 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251358 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251435 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.251480 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.257687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.257721 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.257854 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.262692 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.269475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.323331 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 15:52:37 crc kubenswrapper[4907]: I0313 15:52:37.840357 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz"] Mar 13 15:52:38 crc kubenswrapper[4907]: I0313 15:52:38.250679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerStarted","Data":"4e306d67a326c5a22924c14c80f0301c152e3eb9bef16aa1cc486b8e33a33cc7"} Mar 13 15:52:39 crc kubenswrapper[4907]: I0313 15:52:39.188945 4907 scope.go:117] "RemoveContainer" containerID="971365269f08ef2266842a1c0ca18c6bb47126b67a1360b35d5a7dfd3acab380" Mar 13 15:52:39 crc kubenswrapper[4907]: I0313 15:52:39.264042 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerStarted","Data":"384d76deb615542ee9afab32cbd864c5a2ef0e524a3354232d127d85ac034766"} Mar 13 15:52:39 crc kubenswrapper[4907]: I0313 15:52:39.310276 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" podStartSLOduration=2.524722311 podStartE2EDuration="3.310247263s" podCreationTimestamp="2026-03-13 15:52:36 +0000 UTC" firstStartedPulling="2026-03-13 15:52:37.853966894 +0000 UTC m=+6456.753754583" lastFinishedPulling="2026-03-13 15:52:38.639491846 +0000 UTC m=+6457.539279535" observedRunningTime="2026-03-13 15:52:39.284281743 +0000 UTC m=+6458.184069472" watchObservedRunningTime="2026-03-13 15:52:39.310247263 +0000 UTC m=+6458.210034982" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.041330 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.042088 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.042151 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.043329 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.043418 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" gracePeriod=600 Mar 13 15:52:48 crc kubenswrapper[4907]: E0313 15:52:48.184056 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.350440 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" exitCode=0 Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.350486 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792"} Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.350521 4907 scope.go:117] "RemoveContainer" containerID="81d94e236cbdff3b86e8375b8abbaee97f49d844932ee0151cdfa9d97cb1c3b2" Mar 13 15:52:48 crc kubenswrapper[4907]: I0313 15:52:48.351071 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:52:48 crc kubenswrapper[4907]: E0313 15:52:48.351388 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.626048 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.630248 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.631832 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.714550 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.714839 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.715113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.816779 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.816845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.816919 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.817814 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.817872 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.837143 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"redhat-marketplace-6fnjv\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:54 crc kubenswrapper[4907]: I0313 15:52:54.958255 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:52:55 crc kubenswrapper[4907]: I0313 15:52:55.430644 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:52:55 crc kubenswrapper[4907]: W0313 15:52:55.440497 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63bc7557_0465_42dc_873c_a8eb54e03a45.slice/crio-ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed WatchSource:0}: Error finding container ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed: Status 404 returned error can't find the container with id ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed Mar 13 15:52:56 crc kubenswrapper[4907]: I0313 15:52:56.437869 4907 generic.go:334] "Generic (PLEG): container finished" podID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" exitCode=0 Mar 13 15:52:56 crc kubenswrapper[4907]: I0313 15:52:56.438099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726"} Mar 13 15:52:56 crc kubenswrapper[4907]: I0313 15:52:56.438181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerStarted","Data":"ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed"} Mar 13 15:52:57 crc kubenswrapper[4907]: I0313 15:52:57.452000 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerStarted","Data":"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85"} Mar 13 15:52:58 crc kubenswrapper[4907]: I0313 15:52:58.465419 4907 generic.go:334] "Generic (PLEG): container finished" podID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" exitCode=0 Mar 13 15:52:58 crc kubenswrapper[4907]: I0313 15:52:58.465672 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85"} Mar 13 15:52:59 crc kubenswrapper[4907]: I0313 15:52:59.477421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerStarted","Data":"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229"} Mar 13 15:52:59 crc kubenswrapper[4907]: I0313 15:52:59.506003 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6fnjv" podStartSLOduration=3.013929319 podStartE2EDuration="5.505980356s" podCreationTimestamp="2026-03-13 15:52:54 +0000 UTC" firstStartedPulling="2026-03-13 15:52:56.443332329 +0000 UTC m=+6475.343120018" lastFinishedPulling="2026-03-13 15:52:58.935383366 +0000 UTC m=+6477.835171055" observedRunningTime="2026-03-13 15:52:59.495791508 +0000 UTC m=+6478.395579217" watchObservedRunningTime="2026-03-13 15:52:59.505980356 +0000 UTC m=+6478.405768055" Mar 13 15:52:59 crc kubenswrapper[4907]: I0313 15:52:59.782469 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:52:59 crc kubenswrapper[4907]: E0313 15:52:59.782864 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:04 crc kubenswrapper[4907]: I0313 15:53:04.959102 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:04 crc kubenswrapper[4907]: I0313 15:53:04.959593 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:05 crc kubenswrapper[4907]: I0313 15:53:05.012760 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:05 crc kubenswrapper[4907]: I0313 15:53:05.608354 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:05 crc kubenswrapper[4907]: I0313 15:53:05.663237 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:53:07 crc kubenswrapper[4907]: I0313 15:53:07.575174 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6fnjv" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" containerID="cri-o://7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" gracePeriod=2 Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.046145 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.107155 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") pod \"63bc7557-0465-42dc-873c-a8eb54e03a45\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.107520 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") pod \"63bc7557-0465-42dc-873c-a8eb54e03a45\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.107556 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") pod \"63bc7557-0465-42dc-873c-a8eb54e03a45\" (UID: \"63bc7557-0465-42dc-873c-a8eb54e03a45\") " Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.108311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities" (OuterVolumeSpecName: "utilities") pod "63bc7557-0465-42dc-873c-a8eb54e03a45" (UID: "63bc7557-0465-42dc-873c-a8eb54e03a45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.115212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl" (OuterVolumeSpecName: "kube-api-access-d9qjl") pod "63bc7557-0465-42dc-873c-a8eb54e03a45" (UID: "63bc7557-0465-42dc-873c-a8eb54e03a45"). InnerVolumeSpecName "kube-api-access-d9qjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.139029 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63bc7557-0465-42dc-873c-a8eb54e03a45" (UID: "63bc7557-0465-42dc-873c-a8eb54e03a45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.210045 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.210080 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63bc7557-0465-42dc-873c-a8eb54e03a45-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.210093 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9qjl\" (UniqueName: \"kubernetes.io/projected/63bc7557-0465-42dc-873c-a8eb54e03a45-kube-api-access-d9qjl\") on node \"crc\" DevicePath \"\"" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.586961 4907 generic.go:334] "Generic (PLEG): container finished" podID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" exitCode=0 Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587046 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6fnjv" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229"} Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587171 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6fnjv" event={"ID":"63bc7557-0465-42dc-873c-a8eb54e03a45","Type":"ContainerDied","Data":"ebcd18dfd1b4c9a8759ca1e5be43efea680a80538844ba60fe37c5d9da7366ed"} Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.587196 4907 scope.go:117] "RemoveContainer" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.614443 4907 scope.go:117] "RemoveContainer" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.624384 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.635602 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6fnjv"] Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.642839 4907 scope.go:117] "RemoveContainer" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.684945 4907 scope.go:117] "RemoveContainer" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" Mar 13 15:53:08 crc kubenswrapper[4907]: E0313 15:53:08.685465 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229\": container with ID starting with 7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229 not found: ID does not exist" containerID="7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685506 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229"} err="failed to get container status \"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229\": rpc error: code = NotFound desc = could not find container \"7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229\": container with ID starting with 7a058c3c97b1e6d74dc7c5eefe16beabcf70d2ab01c388a4525e76f026793229 not found: ID does not exist" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685533 4907 scope.go:117] "RemoveContainer" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" Mar 13 15:53:08 crc kubenswrapper[4907]: E0313 15:53:08.685936 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85\": container with ID starting with 321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85 not found: ID does not exist" containerID="321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685971 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85"} err="failed to get container status \"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85\": rpc error: code = NotFound desc = could not find container \"321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85\": container with ID starting with 321294b102a32d438b57983725dd613a9c65287a746583d505990a3f7122ab85 not found: ID does not exist" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.685988 4907 scope.go:117] "RemoveContainer" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" Mar 13 15:53:08 crc kubenswrapper[4907]: E0313 15:53:08.686276 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726\": container with ID starting with 80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726 not found: ID does not exist" containerID="80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726" Mar 13 15:53:08 crc kubenswrapper[4907]: I0313 15:53:08.686304 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726"} err="failed to get container status \"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726\": rpc error: code = NotFound desc = could not find container \"80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726\": container with ID starting with 80503b331ede81d6354f232f75d1fa04ac94047782d538502272582284122726 not found: ID does not exist" Mar 13 15:53:09 crc kubenswrapper[4907]: I0313 15:53:09.796389 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" path="/var/lib/kubelet/pods/63bc7557-0465-42dc-873c-a8eb54e03a45/volumes" Mar 13 15:53:14 crc kubenswrapper[4907]: I0313 15:53:14.783227 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:14 crc kubenswrapper[4907]: E0313 15:53:14.783980 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:25 crc kubenswrapper[4907]: I0313 15:53:25.782344 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:25 crc kubenswrapper[4907]: E0313 15:53:25.783143 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:38 crc kubenswrapper[4907]: I0313 15:53:38.783446 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:38 crc kubenswrapper[4907]: E0313 15:53:38.784174 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:53:53 crc kubenswrapper[4907]: I0313 15:53:53.783648 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:53:53 crc kubenswrapper[4907]: E0313 15:53:53.784374 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.171298 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 15:54:00 crc kubenswrapper[4907]: E0313 15:54:00.173195 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-content" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.173600 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-content" Mar 13 15:54:00 crc kubenswrapper[4907]: E0313 15:54:00.175212 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-utilities" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.175243 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="extract-utilities" Mar 13 15:54:00 crc kubenswrapper[4907]: E0313 15:54:00.175280 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.175290 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.175740 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="63bc7557-0465-42dc-873c-a8eb54e03a45" containerName="registry-server" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.176725 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.178942 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.179077 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.180386 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.184272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.273505 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"auto-csr-approver-29556954-kn875\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.376460 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"auto-csr-approver-29556954-kn875\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.394853 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"auto-csr-approver-29556954-kn875\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.509514 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.842437 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 15:54:00 crc kubenswrapper[4907]: I0313 15:54:00.842656 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 15:54:01 crc kubenswrapper[4907]: I0313 15:54:01.094832 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556954-kn875" event={"ID":"97bac75d-c6b8-4d2e-8a24-d0c6279347b5","Type":"ContainerStarted","Data":"cc96e61a9c12fab292fda588e891fee3cfcbebd049975f4172aaba1347b1eabd"} Mar 13 15:54:02 crc kubenswrapper[4907]: E0313 15:54:02.640627 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97bac75d_c6b8_4d2e_8a24_d0c6279347b5.slice/crio-4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97bac75d_c6b8_4d2e_8a24_d0c6279347b5.slice/crio-conmon-4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:54:03 crc kubenswrapper[4907]: I0313 15:54:03.114038 4907 generic.go:334] "Generic (PLEG): container finished" podID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerID="4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa" exitCode=0 Mar 13 15:54:03 crc kubenswrapper[4907]: I0313 15:54:03.114109 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556954-kn875" event={"ID":"97bac75d-c6b8-4d2e-8a24-d0c6279347b5","Type":"ContainerDied","Data":"4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa"} Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.493463 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.572332 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") pod \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\" (UID: \"97bac75d-c6b8-4d2e-8a24-d0c6279347b5\") " Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.580312 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh" (OuterVolumeSpecName: "kube-api-access-m79fh") pod "97bac75d-c6b8-4d2e-8a24-d0c6279347b5" (UID: "97bac75d-c6b8-4d2e-8a24-d0c6279347b5"). InnerVolumeSpecName "kube-api-access-m79fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:54:04 crc kubenswrapper[4907]: I0313 15:54:04.675730 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m79fh\" (UniqueName: \"kubernetes.io/projected/97bac75d-c6b8-4d2e-8a24-d0c6279347b5-kube-api-access-m79fh\") on node \"crc\" DevicePath \"\"" Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.135683 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556954-kn875" event={"ID":"97bac75d-c6b8-4d2e-8a24-d0c6279347b5","Type":"ContainerDied","Data":"cc96e61a9c12fab292fda588e891fee3cfcbebd049975f4172aaba1347b1eabd"} Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.135722 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc96e61a9c12fab292fda588e891fee3cfcbebd049975f4172aaba1347b1eabd" Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.135771 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556954-kn875" Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.551593 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.560177 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556948-mmprl"] Mar 13 15:54:05 crc kubenswrapper[4907]: I0313 15:54:05.795088 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="673cadb2-82bd-4e26-8222-5a89c310f597" path="/var/lib/kubelet/pods/673cadb2-82bd-4e26-8222-5a89c310f597/volumes" Mar 13 15:54:08 crc kubenswrapper[4907]: I0313 15:54:08.783011 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:08 crc kubenswrapper[4907]: E0313 15:54:08.783855 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:21 crc kubenswrapper[4907]: I0313 15:54:21.788554 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:21 crc kubenswrapper[4907]: E0313 15:54:21.789360 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:24 crc kubenswrapper[4907]: I0313 15:54:24.043395 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:54:24 crc kubenswrapper[4907]: I0313 15:54:24.052372 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-ghczh"] Mar 13 15:54:25 crc kubenswrapper[4907]: I0313 15:54:25.798324 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa6e501-c175-44be-af00-02b8aac76bab" path="/var/lib/kubelet/pods/0aa6e501-c175-44be-af00-02b8aac76bab/volumes" Mar 13 15:54:26 crc kubenswrapper[4907]: I0313 15:54:26.026759 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:54:26 crc kubenswrapper[4907]: I0313 15:54:26.035415 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-6234-account-create-update-wjdx9"] Mar 13 15:54:27 crc kubenswrapper[4907]: I0313 15:54:27.810202 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c420aeff-7a0b-4efa-b9c0-a93791eb0378" path="/var/lib/kubelet/pods/c420aeff-7a0b-4efa-b9c0-a93791eb0378/volumes" Mar 13 15:54:32 crc kubenswrapper[4907]: I0313 15:54:32.027504 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:54:32 crc kubenswrapper[4907]: I0313 15:54:32.037450 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-p9hf2"] Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.030580 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.039453 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-1da8-account-create-update-82kpd"] Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.796128 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b46488e-1fef-479f-94b6-1958e5174219" path="/var/lib/kubelet/pods/3b46488e-1fef-479f-94b6-1958e5174219/volumes" Mar 13 15:54:33 crc kubenswrapper[4907]: I0313 15:54:33.798278 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d218d11d-2afc-4f21-add2-af15cd8c50ad" path="/var/lib/kubelet/pods/d218d11d-2afc-4f21-add2-af15cd8c50ad/volumes" Mar 13 15:54:36 crc kubenswrapper[4907]: I0313 15:54:36.781966 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:36 crc kubenswrapper[4907]: E0313 15:54:36.783445 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.360263 4907 scope.go:117] "RemoveContainer" containerID="1d71acbd6f06389aec2ad567011e53ff359b678fb6ad5c86cc822259fbdd55b9" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.384895 4907 scope.go:117] "RemoveContainer" containerID="ef160b1245008062ab4f675d742f68b73acd696e5db3bbb49ff009cadb324f56" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.436969 4907 scope.go:117] "RemoveContainer" containerID="48d98d79386ad3e2c5200a51cdf22425a3d47af926f18ea01f01295cd4a45c13" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.493430 4907 scope.go:117] "RemoveContainer" containerID="f56058903e847c31c47f45cadeb76ef300892a82a5c55a4fd28b47f0d9e1eeec" Mar 13 15:54:39 crc kubenswrapper[4907]: I0313 15:54:39.534997 4907 scope.go:117] "RemoveContainer" containerID="e63bf49f2a3afc218c4b126f220439d3e3bcaa22e7a77cab733b2ce5385658bb" Mar 13 15:54:51 crc kubenswrapper[4907]: I0313 15:54:51.790019 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:54:51 crc kubenswrapper[4907]: E0313 15:54:51.791042 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:04 crc kubenswrapper[4907]: I0313 15:55:04.782813 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:04 crc kubenswrapper[4907]: E0313 15:55:04.784586 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:17 crc kubenswrapper[4907]: I0313 15:55:17.046382 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:55:17 crc kubenswrapper[4907]: I0313 15:55:17.057513 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-rjfnd"] Mar 13 15:55:17 crc kubenswrapper[4907]: I0313 15:55:17.795010 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aefbf41f-fa5a-4763-af85-3c9e8df33b6a" path="/var/lib/kubelet/pods/aefbf41f-fa5a-4763-af85-3c9e8df33b6a/volumes" Mar 13 15:55:19 crc kubenswrapper[4907]: I0313 15:55:19.782536 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:19 crc kubenswrapper[4907]: E0313 15:55:19.784021 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:32 crc kubenswrapper[4907]: I0313 15:55:32.782925 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:32 crc kubenswrapper[4907]: E0313 15:55:32.783821 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:39 crc kubenswrapper[4907]: I0313 15:55:39.663851 4907 scope.go:117] "RemoveContainer" containerID="94c59567d9e99f3b6037554dff6847cedf9844e341b49a4a37077ea52d3428be" Mar 13 15:55:39 crc kubenswrapper[4907]: I0313 15:55:39.692297 4907 scope.go:117] "RemoveContainer" containerID="4586239eb0dcd4e94af6c8d52a6990fcfc7c67962591f11127b749cbaee23663" Mar 13 15:55:44 crc kubenswrapper[4907]: I0313 15:55:44.783122 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:44 crc kubenswrapper[4907]: E0313 15:55:44.784007 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:55:58 crc kubenswrapper[4907]: I0313 15:55:58.783111 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:55:58 crc kubenswrapper[4907]: E0313 15:55:58.784118 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.158937 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 15:56:00 crc kubenswrapper[4907]: E0313 15:56:00.159865 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerName="oc" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.159902 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerName="oc" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.160189 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" containerName="oc" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.161213 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.163431 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.163726 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.167259 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"auto-csr-approver-29556956-txd4m\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.167487 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.184870 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.269436 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"auto-csr-approver-29556956-txd4m\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.291124 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"auto-csr-approver-29556956-txd4m\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.488583 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:00 crc kubenswrapper[4907]: I0313 15:56:00.970362 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 15:56:01 crc kubenswrapper[4907]: I0313 15:56:01.252657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556956-txd4m" event={"ID":"a22eedb1-70a8-4232-9a54-8625a4dd7022","Type":"ContainerStarted","Data":"2493883c79e986e8d864cbb5316247394c34dc9c3fd5ce98c48828fdd9942fc7"} Mar 13 15:56:03 crc kubenswrapper[4907]: I0313 15:56:03.279491 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556956-txd4m" event={"ID":"a22eedb1-70a8-4232-9a54-8625a4dd7022","Type":"ContainerDied","Data":"76fd8f67a17de0068ab3b4548c472b268fc93309cdb749646d71dbe2fb8aef2e"} Mar 13 15:56:03 crc kubenswrapper[4907]: I0313 15:56:03.279507 4907 generic.go:334] "Generic (PLEG): container finished" podID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerID="76fd8f67a17de0068ab3b4548c472b268fc93309cdb749646d71dbe2fb8aef2e" exitCode=0 Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.681166 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.884136 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") pod \"a22eedb1-70a8-4232-9a54-8625a4dd7022\" (UID: \"a22eedb1-70a8-4232-9a54-8625a4dd7022\") " Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.893228 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx" (OuterVolumeSpecName: "kube-api-access-rg9kx") pod "a22eedb1-70a8-4232-9a54-8625a4dd7022" (UID: "a22eedb1-70a8-4232-9a54-8625a4dd7022"). InnerVolumeSpecName "kube-api-access-rg9kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:56:04 crc kubenswrapper[4907]: I0313 15:56:04.987927 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rg9kx\" (UniqueName: \"kubernetes.io/projected/a22eedb1-70a8-4232-9a54-8625a4dd7022-kube-api-access-rg9kx\") on node \"crc\" DevicePath \"\"" Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.303713 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556956-txd4m" event={"ID":"a22eedb1-70a8-4232-9a54-8625a4dd7022","Type":"ContainerDied","Data":"2493883c79e986e8d864cbb5316247394c34dc9c3fd5ce98c48828fdd9942fc7"} Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.304335 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2493883c79e986e8d864cbb5316247394c34dc9c3fd5ce98c48828fdd9942fc7" Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.303767 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556956-txd4m" Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.774942 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:56:05 crc kubenswrapper[4907]: I0313 15:56:05.798431 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556950-pqwgk"] Mar 13 15:56:07 crc kubenswrapper[4907]: I0313 15:56:07.795675 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63d124ed-9ee2-437b-8ebd-b38a53e499af" path="/var/lib/kubelet/pods/63d124ed-9ee2-437b-8ebd-b38a53e499af/volumes" Mar 13 15:56:10 crc kubenswrapper[4907]: I0313 15:56:10.783684 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:10 crc kubenswrapper[4907]: E0313 15:56:10.785369 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:23 crc kubenswrapper[4907]: I0313 15:56:23.782640 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:23 crc kubenswrapper[4907]: E0313 15:56:23.783323 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:37 crc kubenswrapper[4907]: I0313 15:56:37.782754 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:37 crc kubenswrapper[4907]: E0313 15:56:37.783585 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:56:39 crc kubenswrapper[4907]: I0313 15:56:39.792565 4907 scope.go:117] "RemoveContainer" containerID="fa4769654b9ebff08198d548ba01872a5dbbbcb75936088b49f7d253d89a76c0" Mar 13 15:56:49 crc kubenswrapper[4907]: I0313 15:56:49.783820 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:56:49 crc kubenswrapper[4907]: E0313 15:56:49.784554 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:02 crc kubenswrapper[4907]: I0313 15:57:02.783519 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:02 crc kubenswrapper[4907]: E0313 15:57:02.786387 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.062843 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.080330 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.089104 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-lnsws"] Mar 13 15:57:14 crc kubenswrapper[4907]: I0313 15:57:14.096960 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-019b-account-create-update-848sw"] Mar 13 15:57:15 crc kubenswrapper[4907]: I0313 15:57:15.796459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="490bd531-09f6-48e5-b194-b79e9b532131" path="/var/lib/kubelet/pods/490bd531-09f6-48e5-b194-b79e9b532131/volumes" Mar 13 15:57:15 crc kubenswrapper[4907]: I0313 15:57:15.797463 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce0b08b9-c4e1-4236-8f54-cf78b45d2d49" path="/var/lib/kubelet/pods/ce0b08b9-c4e1-4236-8f54-cf78b45d2d49/volumes" Mar 13 15:57:17 crc kubenswrapper[4907]: I0313 15:57:17.782306 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:17 crc kubenswrapper[4907]: E0313 15:57:17.782826 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:30 crc kubenswrapper[4907]: I0313 15:57:30.038118 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:57:30 crc kubenswrapper[4907]: I0313 15:57:30.054117 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-mqnk8"] Mar 13 15:57:30 crc kubenswrapper[4907]: I0313 15:57:30.782589 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:30 crc kubenswrapper[4907]: E0313 15:57:30.782849 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:31 crc kubenswrapper[4907]: I0313 15:57:31.793174 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6df151b5-1c14-488c-9f23-9d6d629b493d" path="/var/lib/kubelet/pods/6df151b5-1c14-488c-9f23-9d6d629b493d/volumes" Mar 13 15:57:39 crc kubenswrapper[4907]: I0313 15:57:39.879768 4907 scope.go:117] "RemoveContainer" containerID="daa82cb23c63e51c880db96744e0980e9ddf6f38f0cbcf2713496cba5c2d4486" Mar 13 15:57:39 crc kubenswrapper[4907]: I0313 15:57:39.913422 4907 scope.go:117] "RemoveContainer" containerID="839f8683996fb111c90ced80fc315df41501ccd29aa3ce62746269e8e8d6d880" Mar 13 15:57:39 crc kubenswrapper[4907]: I0313 15:57:39.959639 4907 scope.go:117] "RemoveContainer" containerID="28274d3b65f2c9d6e1b8e4c19babc239c1199e687ba7bf22cb4ab37476252833" Mar 13 15:57:41 crc kubenswrapper[4907]: I0313 15:57:41.794786 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:41 crc kubenswrapper[4907]: E0313 15:57:41.795247 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 15:57:52 crc kubenswrapper[4907]: I0313 15:57:52.782299 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 15:57:53 crc kubenswrapper[4907]: I0313 15:57:53.350619 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19"} Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.152207 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 15:58:00 crc kubenswrapper[4907]: E0313 15:58:00.154683 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerName="oc" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.154814 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerName="oc" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.155312 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" containerName="oc" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.156654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.159076 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.159716 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.160136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.167185 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.314114 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"auto-csr-approver-29556958-dk8dz\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.416202 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"auto-csr-approver-29556958-dk8dz\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.437861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"auto-csr-approver-29556958-dk8dz\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.490162 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:00 crc kubenswrapper[4907]: I0313 15:58:00.962665 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 15:58:01 crc kubenswrapper[4907]: I0313 15:58:01.427571 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerStarted","Data":"bb6038124e3db92e462b38fae6dbf334e4cb83a55a475138ec56e8d71be7fae6"} Mar 13 15:58:02 crc kubenswrapper[4907]: I0313 15:58:02.439170 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerStarted","Data":"d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7"} Mar 13 15:58:02 crc kubenswrapper[4907]: I0313 15:58:02.460515 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" podStartSLOduration=1.320104216 podStartE2EDuration="2.460496856s" podCreationTimestamp="2026-03-13 15:58:00 +0000 UTC" firstStartedPulling="2026-03-13 15:58:00.966859772 +0000 UTC m=+6779.866647451" lastFinishedPulling="2026-03-13 15:58:02.107252382 +0000 UTC m=+6781.007040091" observedRunningTime="2026-03-13 15:58:02.450678037 +0000 UTC m=+6781.350465726" watchObservedRunningTime="2026-03-13 15:58:02.460496856 +0000 UTC m=+6781.360284545" Mar 13 15:58:03 crc kubenswrapper[4907]: I0313 15:58:03.449041 4907 generic.go:334] "Generic (PLEG): container finished" podID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerID="d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7" exitCode=0 Mar 13 15:58:03 crc kubenswrapper[4907]: I0313 15:58:03.449150 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerDied","Data":"d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7"} Mar 13 15:58:04 crc kubenswrapper[4907]: I0313 15:58:04.828297 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:04 crc kubenswrapper[4907]: I0313 15:58:04.911970 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") pod \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\" (UID: \"2d38cd8e-5a08-4c56-b35e-88617df1d3a8\") " Mar 13 15:58:04 crc kubenswrapper[4907]: I0313 15:58:04.920388 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8" (OuterVolumeSpecName: "kube-api-access-4ncj8") pod "2d38cd8e-5a08-4c56-b35e-88617df1d3a8" (UID: "2d38cd8e-5a08-4c56-b35e-88617df1d3a8"). InnerVolumeSpecName "kube-api-access-4ncj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.014754 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ncj8\" (UniqueName: \"kubernetes.io/projected/2d38cd8e-5a08-4c56-b35e-88617df1d3a8-kube-api-access-4ncj8\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.469814 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" event={"ID":"2d38cd8e-5a08-4c56-b35e-88617df1d3a8","Type":"ContainerDied","Data":"bb6038124e3db92e462b38fae6dbf334e4cb83a55a475138ec56e8d71be7fae6"} Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.469852 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb6038124e3db92e462b38fae6dbf334e4cb83a55a475138ec56e8d71be7fae6" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.469869 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556958-dk8dz" Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.903818 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:58:05 crc kubenswrapper[4907]: I0313 15:58:05.912515 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556952-b69cd"] Mar 13 15:58:07 crc kubenswrapper[4907]: I0313 15:58:07.794448 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2915500-b9a0-45a4-960e-5325f53b445c" path="/var/lib/kubelet/pods/d2915500-b9a0-45a4-960e-5325f53b445c/volumes" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.421834 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:16 crc kubenswrapper[4907]: E0313 15:58:16.422929 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerName="oc" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.422946 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerName="oc" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.423232 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" containerName="oc" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.425182 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.439117 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.583204 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.583291 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.583595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.685822 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.685972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.686107 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.686352 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.686599 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.707435 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"community-operators-pk7rd\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:16 crc kubenswrapper[4907]: I0313 15:58:16.803663 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:17 crc kubenswrapper[4907]: I0313 15:58:17.387500 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:17 crc kubenswrapper[4907]: I0313 15:58:17.611469 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080"} Mar 13 15:58:17 crc kubenswrapper[4907]: I0313 15:58:17.611531 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"31bb1f1b4403f006239e8d040d3abbb5c7a93523d9ec4c3f298d90363dac8352"} Mar 13 15:58:18 crc kubenswrapper[4907]: I0313 15:58:18.622465 4907 generic.go:334] "Generic (PLEG): container finished" podID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" exitCode=0 Mar 13 15:58:18 crc kubenswrapper[4907]: I0313 15:58:18.622657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080"} Mar 13 15:58:19 crc kubenswrapper[4907]: I0313 15:58:19.632262 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702"} Mar 13 15:58:21 crc kubenswrapper[4907]: I0313 15:58:21.651818 4907 generic.go:334] "Generic (PLEG): container finished" podID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" exitCode=0 Mar 13 15:58:21 crc kubenswrapper[4907]: I0313 15:58:21.651916 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702"} Mar 13 15:58:22 crc kubenswrapper[4907]: I0313 15:58:22.663144 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerStarted","Data":"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07"} Mar 13 15:58:22 crc kubenswrapper[4907]: I0313 15:58:22.680565 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pk7rd" podStartSLOduration=2.931828724 podStartE2EDuration="6.680548074s" podCreationTimestamp="2026-03-13 15:58:16 +0000 UTC" firstStartedPulling="2026-03-13 15:58:18.6263783 +0000 UTC m=+6797.526165989" lastFinishedPulling="2026-03-13 15:58:22.37509764 +0000 UTC m=+6801.274885339" observedRunningTime="2026-03-13 15:58:22.677209833 +0000 UTC m=+6801.576997542" watchObservedRunningTime="2026-03-13 15:58:22.680548074 +0000 UTC m=+6801.580335773" Mar 13 15:58:26 crc kubenswrapper[4907]: I0313 15:58:26.803928 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:26 crc kubenswrapper[4907]: I0313 15:58:26.804580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:26 crc kubenswrapper[4907]: I0313 15:58:26.853282 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:27 crc kubenswrapper[4907]: I0313 15:58:27.764628 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:27 crc kubenswrapper[4907]: I0313 15:58:27.816603 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:29 crc kubenswrapper[4907]: I0313 15:58:29.737390 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pk7rd" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" containerID="cri-o://ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" gracePeriod=2 Mar 13 15:58:29 crc kubenswrapper[4907]: E0313 15:58:29.799798 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac68b7dd_54d2_4691_ac90_4f6f5ff0bc43.slice/crio-ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07.scope\": RecentStats: unable to find data in memory cache]" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.240668 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.400587 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") pod \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.400840 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") pod \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.400917 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") pod \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\" (UID: \"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43\") " Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.402386 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities" (OuterVolumeSpecName: "utilities") pod "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" (UID: "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.406080 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc" (OuterVolumeSpecName: "kube-api-access-rrgtc") pod "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" (UID: "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43"). InnerVolumeSpecName "kube-api-access-rrgtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.469679 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" (UID: "ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.503495 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.503528 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrgtc\" (UniqueName: \"kubernetes.io/projected/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-kube-api-access-rrgtc\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.503539 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751218 4907 generic.go:334] "Generic (PLEG): container finished" podID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" exitCode=0 Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751259 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07"} Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751285 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pk7rd" event={"ID":"ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43","Type":"ContainerDied","Data":"31bb1f1b4403f006239e8d040d3abbb5c7a93523d9ec4c3f298d90363dac8352"} Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751303 4907 scope.go:117] "RemoveContainer" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.751352 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pk7rd" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.793735 4907 scope.go:117] "RemoveContainer" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.793891 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.816553 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pk7rd"] Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.830345 4907 scope.go:117] "RemoveContainer" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.905192 4907 scope.go:117] "RemoveContainer" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" Mar 13 15:58:30 crc kubenswrapper[4907]: E0313 15:58:30.906130 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07\": container with ID starting with ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07 not found: ID does not exist" containerID="ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906176 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07"} err="failed to get container status \"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07\": rpc error: code = NotFound desc = could not find container \"ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07\": container with ID starting with ffa10747043fef78a41d5b63379aa594a58eda0c580b0fdb3ed18e4cc5b00c07 not found: ID does not exist" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906201 4907 scope.go:117] "RemoveContainer" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" Mar 13 15:58:30 crc kubenswrapper[4907]: E0313 15:58:30.906693 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702\": container with ID starting with db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702 not found: ID does not exist" containerID="db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906730 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702"} err="failed to get container status \"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702\": rpc error: code = NotFound desc = could not find container \"db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702\": container with ID starting with db5c85295b78a586f9a9d79869cb88866e67ead5e0fb9606d34998d9455a4702 not found: ID does not exist" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.906751 4907 scope.go:117] "RemoveContainer" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" Mar 13 15:58:30 crc kubenswrapper[4907]: E0313 15:58:30.907191 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080\": container with ID starting with 53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080 not found: ID does not exist" containerID="53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080" Mar 13 15:58:30 crc kubenswrapper[4907]: I0313 15:58:30.907224 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080"} err="failed to get container status \"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080\": rpc error: code = NotFound desc = could not find container \"53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080\": container with ID starting with 53b83ec7d94b2543162fe15e2a0e9e0d2a45294fa4e4b0c3681b0c95abfa7080 not found: ID does not exist" Mar 13 15:58:31 crc kubenswrapper[4907]: I0313 15:58:31.804384 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" path="/var/lib/kubelet/pods/ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43/volumes" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.304689 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:58:38 crc kubenswrapper[4907]: E0313 15:58:38.305538 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-content" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305555 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-content" Mar 13 15:58:38 crc kubenswrapper[4907]: E0313 15:58:38.305599 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305608 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" Mar 13 15:58:38 crc kubenswrapper[4907]: E0313 15:58:38.305625 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-utilities" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305631 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="extract-utilities" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.305898 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac68b7dd-54d2-4691-ac90-4f6f5ff0bc43" containerName="registry-server" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.307797 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.329257 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.416173 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.416222 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.416340 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.518697 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.518750 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.518947 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.519474 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.519606 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.545802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"redhat-operators-kvhlf\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:38 crc kubenswrapper[4907]: I0313 15:58:38.648704 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.196775 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.853275 4907 generic.go:334] "Generic (PLEG): container finished" podID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" exitCode=0 Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.853530 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978"} Mar 13 15:58:39 crc kubenswrapper[4907]: I0313 15:58:39.853555 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerStarted","Data":"a8505ebf077b55d207686198c5233286e1d24c14f8a53397bbeaddc40331ff03"} Mar 13 15:58:40 crc kubenswrapper[4907]: I0313 15:58:40.091012 4907 scope.go:117] "RemoveContainer" containerID="ae08e1f5613f0f2169631235f70419712208d7aa173e1b4d3da70be8496080b9" Mar 13 15:58:40 crc kubenswrapper[4907]: I0313 15:58:40.866025 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerStarted","Data":"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35"} Mar 13 15:58:45 crc kubenswrapper[4907]: I0313 15:58:45.936691 4907 generic.go:334] "Generic (PLEG): container finished" podID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" exitCode=0 Mar 13 15:58:45 crc kubenswrapper[4907]: I0313 15:58:45.936743 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35"} Mar 13 15:58:46 crc kubenswrapper[4907]: I0313 15:58:46.951062 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerStarted","Data":"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70"} Mar 13 15:58:46 crc kubenswrapper[4907]: I0313 15:58:46.986928 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kvhlf" podStartSLOduration=2.479931635 podStartE2EDuration="8.986900599s" podCreationTimestamp="2026-03-13 15:58:38 +0000 UTC" firstStartedPulling="2026-03-13 15:58:39.855469184 +0000 UTC m=+6818.755256873" lastFinishedPulling="2026-03-13 15:58:46.362438138 +0000 UTC m=+6825.262225837" observedRunningTime="2026-03-13 15:58:46.972177485 +0000 UTC m=+6825.871965184" watchObservedRunningTime="2026-03-13 15:58:46.986900599 +0000 UTC m=+6825.886688318" Mar 13 15:58:48 crc kubenswrapper[4907]: I0313 15:58:48.649051 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:48 crc kubenswrapper[4907]: I0313 15:58:48.649374 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:49 crc kubenswrapper[4907]: I0313 15:58:49.702084 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kvhlf" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" probeResult="failure" output=< Mar 13 15:58:49 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 15:58:49 crc kubenswrapper[4907]: > Mar 13 15:58:58 crc kubenswrapper[4907]: I0313 15:58:58.703947 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:58 crc kubenswrapper[4907]: I0313 15:58:58.757623 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:58:58 crc kubenswrapper[4907]: I0313 15:58:58.944054 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.103029 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kvhlf" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" containerID="cri-o://bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" gracePeriod=2 Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.713529 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.873221 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") pod \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.873519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") pod \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.873610 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") pod \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\" (UID: \"9739d592-69ae-4c44-ba8a-f4a66d3dd37e\") " Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.874311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities" (OuterVolumeSpecName: "utilities") pod "9739d592-69ae-4c44-ba8a-f4a66d3dd37e" (UID: "9739d592-69ae-4c44-ba8a-f4a66d3dd37e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.875542 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.880831 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j" (OuterVolumeSpecName: "kube-api-access-l5p4j") pod "9739d592-69ae-4c44-ba8a-f4a66d3dd37e" (UID: "9739d592-69ae-4c44-ba8a-f4a66d3dd37e"). InnerVolumeSpecName "kube-api-access-l5p4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 15:59:00 crc kubenswrapper[4907]: I0313 15:59:00.977774 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5p4j\" (UniqueName: \"kubernetes.io/projected/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-kube-api-access-l5p4j\") on node \"crc\" DevicePath \"\"" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.026358 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9739d592-69ae-4c44-ba8a-f4a66d3dd37e" (UID: "9739d592-69ae-4c44-ba8a-f4a66d3dd37e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.079946 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9739d592-69ae-4c44-ba8a-f4a66d3dd37e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113849 4907 generic.go:334] "Generic (PLEG): container finished" podID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" exitCode=0 Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113919 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70"} Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113951 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvhlf" event={"ID":"9739d592-69ae-4c44-ba8a-f4a66d3dd37e","Type":"ContainerDied","Data":"a8505ebf077b55d207686198c5233286e1d24c14f8a53397bbeaddc40331ff03"} Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.113973 4907 scope.go:117] "RemoveContainer" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.115454 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvhlf" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.154011 4907 scope.go:117] "RemoveContainer" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.162505 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.178869 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kvhlf"] Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.187209 4907 scope.go:117] "RemoveContainer" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.245184 4907 scope.go:117] "RemoveContainer" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" Mar 13 15:59:01 crc kubenswrapper[4907]: E0313 15:59:01.246053 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70\": container with ID starting with bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70 not found: ID does not exist" containerID="bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.246105 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70"} err="failed to get container status \"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70\": rpc error: code = NotFound desc = could not find container \"bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70\": container with ID starting with bf318b47b118afa14ad5c1f8dd0ece551b6717fd4d445868fc8150982fe2ef70 not found: ID does not exist" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.246140 4907 scope.go:117] "RemoveContainer" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" Mar 13 15:59:01 crc kubenswrapper[4907]: E0313 15:59:01.246934 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35\": container with ID starting with 72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35 not found: ID does not exist" containerID="72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.247004 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35"} err="failed to get container status \"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35\": rpc error: code = NotFound desc = could not find container \"72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35\": container with ID starting with 72383279f5aa6921bbcd47b4d1969465c8d8b98982eda484eee6233bb1706a35 not found: ID does not exist" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.247046 4907 scope.go:117] "RemoveContainer" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" Mar 13 15:59:01 crc kubenswrapper[4907]: E0313 15:59:01.247709 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978\": container with ID starting with fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978 not found: ID does not exist" containerID="fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.247755 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978"} err="failed to get container status \"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978\": rpc error: code = NotFound desc = could not find container \"fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978\": container with ID starting with fe843f7ce1ebfd01e87874065c8fa24f6b69455e96ff16a1e7272446b070a978 not found: ID does not exist" Mar 13 15:59:01 crc kubenswrapper[4907]: I0313 15:59:01.797841 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" path="/var/lib/kubelet/pods/9739d592-69ae-4c44-ba8a-f4a66d3dd37e/volumes" Mar 13 15:59:37 crc kubenswrapper[4907]: I0313 15:59:37.037832 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:59:37 crc kubenswrapper[4907]: I0313 15:59:37.047144 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-6qbgl"] Mar 13 15:59:37 crc kubenswrapper[4907]: I0313 15:59:37.799713 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="689fb651-0113-48ba-b6e4-29c41c5fa21f" path="/var/lib/kubelet/pods/689fb651-0113-48ba-b6e4-29c41c5fa21f/volumes" Mar 13 15:59:38 crc kubenswrapper[4907]: I0313 15:59:38.044165 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:59:38 crc kubenswrapper[4907]: I0313 15:59:38.056497 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-cc97-account-create-update-fc5r7"] Mar 13 15:59:39 crc kubenswrapper[4907]: I0313 15:59:39.810569 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="442daa10-2973-4304-9e1c-039d35226686" path="/var/lib/kubelet/pods/442daa10-2973-4304-9e1c-039d35226686/volumes" Mar 13 15:59:40 crc kubenswrapper[4907]: I0313 15:59:40.177188 4907 scope.go:117] "RemoveContainer" containerID="d38dedf8f9062f909f22038364138fce749f1c07518101c6bfdc3e1d5f365369" Mar 13 15:59:40 crc kubenswrapper[4907]: I0313 15:59:40.218277 4907 scope.go:117] "RemoveContainer" containerID="7eeac1862f5788609b87b66dfacb73f343af69811adf4dfc0c22454194586bf3" Mar 13 15:59:50 crc kubenswrapper[4907]: I0313 15:59:50.050184 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:59:50 crc kubenswrapper[4907]: I0313 15:59:50.066374 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-nwwc8"] Mar 13 15:59:51 crc kubenswrapper[4907]: I0313 15:59:51.807741 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5a77890-eb92-4d55-acf8-481f70c9277c" path="/var/lib/kubelet/pods/c5a77890-eb92-4d55-acf8-481f70c9277c/volumes" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.153034 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:00:00 crc kubenswrapper[4907]: E0313 16:00:00.154055 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154068 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" Mar 13 16:00:00 crc kubenswrapper[4907]: E0313 16:00:00.154088 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-content" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154094 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-content" Mar 13 16:00:00 crc kubenswrapper[4907]: E0313 16:00:00.154114 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-utilities" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154121 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="extract-utilities" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.154325 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9739d592-69ae-4c44-ba8a-f4a66d3dd37e" containerName="registry-server" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.155108 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.157625 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.159099 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.159210 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.163028 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.230157 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"auto-csr-approver-29556960-8xc84\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.250463 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.252709 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.255020 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.255368 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.265779 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.331959 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.332102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.332139 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.332196 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"auto-csr-approver-29556960-8xc84\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.350704 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"auto-csr-approver-29556960-8xc84\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.433468 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.433581 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.434375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.435457 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.436869 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.448209 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"collect-profiles-29556960-mkbxd\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.480090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.572750 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.945657 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:00:00 crc kubenswrapper[4907]: I0313 16:00:00.948964 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:00:01 crc kubenswrapper[4907]: W0313 16:00:01.060481 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1001d4c_6984_4a3c_8d1f_167a6a48ea8a.slice/crio-961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a WatchSource:0}: Error finding container 961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a: Status 404 returned error can't find the container with id 961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.068232 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.796562 4907 generic.go:334] "Generic (PLEG): container finished" podID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerID="0c0e8b2fe415ff36180c0f0cbc79e145510c4e8b607c452be50ff3ed2394d856" exitCode=0 Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.796641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" event={"ID":"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a","Type":"ContainerDied","Data":"0c0e8b2fe415ff36180c0f0cbc79e145510c4e8b607c452be50ff3ed2394d856"} Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.796864 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" event={"ID":"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a","Type":"ContainerStarted","Data":"961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a"} Mar 13 16:00:01 crc kubenswrapper[4907]: I0313 16:00:01.797972 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556960-8xc84" event={"ID":"acb3140e-a115-443a-a39e-5f180a196b13","Type":"ContainerStarted","Data":"4a57c716b0e24bb904434f950bb9538fc19745e51f498ae299351e13dbb5f5e4"} Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.179738 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.294935 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") pod \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.295028 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") pod \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.295073 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") pod \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\" (UID: \"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a\") " Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.295735 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume" (OuterVolumeSpecName: "config-volume") pod "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" (UID: "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.303168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" (UID: "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.303177 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6" (OuterVolumeSpecName: "kube-api-access-mpbb6") pod "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" (UID: "e1001d4c-6984-4a3c-8d1f-167a6a48ea8a"). InnerVolumeSpecName "kube-api-access-mpbb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.398731 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.398769 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpbb6\" (UniqueName: \"kubernetes.io/projected/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-kube-api-access-mpbb6\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.398788 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.821587 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" event={"ID":"e1001d4c-6984-4a3c-8d1f-167a6a48ea8a","Type":"ContainerDied","Data":"961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a"} Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.821640 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="961c155ac80d225a208f0f959fac939f4afb3a793830f43150c57ec2c6d02d2a" Mar 13 16:00:03 crc kubenswrapper[4907]: I0313 16:00:03.821718 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd" Mar 13 16:00:04 crc kubenswrapper[4907]: I0313 16:00:04.248316 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 16:00:04 crc kubenswrapper[4907]: I0313 16:00:04.260686 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556915-jtzgq"] Mar 13 16:00:05 crc kubenswrapper[4907]: I0313 16:00:05.799079 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb6a3c2-e8df-47a4-9ae7-2ee38842051a" path="/var/lib/kubelet/pods/1fb6a3c2-e8df-47a4-9ae7-2ee38842051a/volumes" Mar 13 16:00:10 crc kubenswrapper[4907]: I0313 16:00:10.890458 4907 generic.go:334] "Generic (PLEG): container finished" podID="acb3140e-a115-443a-a39e-5f180a196b13" containerID="90c713255623c6a0cd4f0904c875e6fe6995c5ac18245db4d00a77bd7e6d5d8c" exitCode=0 Mar 13 16:00:10 crc kubenswrapper[4907]: I0313 16:00:10.890615 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556960-8xc84" event={"ID":"acb3140e-a115-443a-a39e-5f180a196b13","Type":"ContainerDied","Data":"90c713255623c6a0cd4f0904c875e6fe6995c5ac18245db4d00a77bd7e6d5d8c"} Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.357506 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.508047 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") pod \"acb3140e-a115-443a-a39e-5f180a196b13\" (UID: \"acb3140e-a115-443a-a39e-5f180a196b13\") " Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.524205 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq" (OuterVolumeSpecName: "kube-api-access-xgvkq") pod "acb3140e-a115-443a-a39e-5f180a196b13" (UID: "acb3140e-a115-443a-a39e-5f180a196b13"). InnerVolumeSpecName "kube-api-access-xgvkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.610865 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgvkq\" (UniqueName: \"kubernetes.io/projected/acb3140e-a115-443a-a39e-5f180a196b13-kube-api-access-xgvkq\") on node \"crc\" DevicePath \"\"" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.916438 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556960-8xc84" event={"ID":"acb3140e-a115-443a-a39e-5f180a196b13","Type":"ContainerDied","Data":"4a57c716b0e24bb904434f950bb9538fc19745e51f498ae299351e13dbb5f5e4"} Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.916789 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a57c716b0e24bb904434f950bb9538fc19745e51f498ae299351e13dbb5f5e4" Mar 13 16:00:12 crc kubenswrapper[4907]: I0313 16:00:12.916524 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556960-8xc84" Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.066971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.083971 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.095683 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-e749-account-create-update-gsxrn"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.106725 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-mzct7"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.417943 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.426550 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556954-kn875"] Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.796576 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0498e113-86c3-4718-a580-13bfdb277ff2" path="/var/lib/kubelet/pods/0498e113-86c3-4718-a580-13bfdb277ff2/volumes" Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.797472 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97bac75d-c6b8-4d2e-8a24-d0c6279347b5" path="/var/lib/kubelet/pods/97bac75d-c6b8-4d2e-8a24-d0c6279347b5/volumes" Mar 13 16:00:13 crc kubenswrapper[4907]: I0313 16:00:13.798281 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a34745-a555-4029-a203-ee17762929e2" path="/var/lib/kubelet/pods/b2a34745-a555-4029-a203-ee17762929e2/volumes" Mar 13 16:00:18 crc kubenswrapper[4907]: I0313 16:00:18.042213 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:00:18 crc kubenswrapper[4907]: I0313 16:00:18.042830 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:00:25 crc kubenswrapper[4907]: I0313 16:00:25.042058 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 16:00:25 crc kubenswrapper[4907]: I0313 16:00:25.055193 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-nhx6q"] Mar 13 16:00:25 crc kubenswrapper[4907]: I0313 16:00:25.793960 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2df4eaaf-eb5c-4e04-b987-5aedb6561f7e" path="/var/lib/kubelet/pods/2df4eaaf-eb5c-4e04-b987-5aedb6561f7e/volumes" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.355303 4907 scope.go:117] "RemoveContainer" containerID="7fe743a125369b9e5af6f2cb7a4bc8d2add93450e84c6dadc626912538eac565" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.380870 4907 scope.go:117] "RemoveContainer" containerID="4dac6251d0188e6b289ee01e6cd109fb2dacb259e342a008dc8836c090408baa" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.453159 4907 scope.go:117] "RemoveContainer" containerID="f8d1778bda06ccfd24371ed557383017005b23e6277f212ccf01bbaf98e64b94" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.514868 4907 scope.go:117] "RemoveContainer" containerID="3da3591c8cb3c5a70194cd0e1a361e544712615fd8402b78ddd2f90160064c36" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.539444 4907 scope.go:117] "RemoveContainer" containerID="8cc12017f30ada4d93b2676752544b7d25e61944ae3bcbe64687e02b090ffb18" Mar 13 16:00:40 crc kubenswrapper[4907]: I0313 16:00:40.592031 4907 scope.go:117] "RemoveContainer" containerID="84256d7f66a6d2a7c62b2540cda3765257adc44483f0deff6d7044e7a488955a" Mar 13 16:00:48 crc kubenswrapper[4907]: I0313 16:00:48.041666 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:00:48 crc kubenswrapper[4907]: I0313 16:00:48.042563 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.756594 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:00:52 crc kubenswrapper[4907]: E0313 16:00:52.757698 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerName="collect-profiles" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.757714 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerName="collect-profiles" Mar 13 16:00:52 crc kubenswrapper[4907]: E0313 16:00:52.757734 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acb3140e-a115-443a-a39e-5f180a196b13" containerName="oc" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.757742 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="acb3140e-a115-443a-a39e-5f180a196b13" containerName="oc" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.758005 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="acb3140e-a115-443a-a39e-5f180a196b13" containerName="oc" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.758021 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" containerName="collect-profiles" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.760110 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.777687 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.952615 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.954081 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:52 crc kubenswrapper[4907]: I0313 16:00:52.954263 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.057481 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.057601 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.057635 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.058145 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.058223 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.078051 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"certified-operators-7nbgp\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.115688 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:00:53 crc kubenswrapper[4907]: I0313 16:00:53.664369 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:00:54 crc kubenswrapper[4907]: I0313 16:00:54.353802 4907 generic.go:334] "Generic (PLEG): container finished" podID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" exitCode=0 Mar 13 16:00:54 crc kubenswrapper[4907]: I0313 16:00:54.353945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41"} Mar 13 16:00:54 crc kubenswrapper[4907]: I0313 16:00:54.354150 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerStarted","Data":"673546a70a0ddd984900b7492ee18632405c8a55ae70039f43b9cdb5ae87218b"} Mar 13 16:00:56 crc kubenswrapper[4907]: I0313 16:00:56.387956 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerStarted","Data":"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213"} Mar 13 16:00:57 crc kubenswrapper[4907]: I0313 16:00:57.401933 4907 generic.go:334] "Generic (PLEG): container finished" podID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" exitCode=0 Mar 13 16:00:57 crc kubenswrapper[4907]: I0313 16:00:57.402009 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213"} Mar 13 16:00:59 crc kubenswrapper[4907]: I0313 16:00:59.432600 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerStarted","Data":"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480"} Mar 13 16:00:59 crc kubenswrapper[4907]: I0313 16:00:59.468757 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7nbgp" podStartSLOduration=3.459966921 podStartE2EDuration="7.468733582s" podCreationTimestamp="2026-03-13 16:00:52 +0000 UTC" firstStartedPulling="2026-03-13 16:00:54.355803613 +0000 UTC m=+6953.255591302" lastFinishedPulling="2026-03-13 16:00:58.364570244 +0000 UTC m=+6957.264357963" observedRunningTime="2026-03-13 16:00:59.45369479 +0000 UTC m=+6958.353482519" watchObservedRunningTime="2026-03-13 16:00:59.468733582 +0000 UTC m=+6958.368521281" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.168039 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29556961-gshbz"] Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.170956 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.186096 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556961-gshbz"] Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335130 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335181 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.335673 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437428 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437625 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437744 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.437777 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.443791 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.448673 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.450173 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.460065 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"keystone-cron-29556961-gshbz\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.505490 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:00 crc kubenswrapper[4907]: I0313 16:01:00.981168 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556961-gshbz"] Mar 13 16:01:00 crc kubenswrapper[4907]: W0313 16:01:00.984457 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode548a313_fc14_4ce9_b268_bf37c79eed83.slice/crio-b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1 WatchSource:0}: Error finding container b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1: Status 404 returned error can't find the container with id b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1 Mar 13 16:01:01 crc kubenswrapper[4907]: I0313 16:01:01.453249 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerStarted","Data":"4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979"} Mar 13 16:01:01 crc kubenswrapper[4907]: I0313 16:01:01.453327 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerStarted","Data":"b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1"} Mar 13 16:01:01 crc kubenswrapper[4907]: I0313 16:01:01.486058 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29556961-gshbz" podStartSLOduration=1.4860244360000001 podStartE2EDuration="1.486024436s" podCreationTimestamp="2026-03-13 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:01:01.480968997 +0000 UTC m=+6960.380756686" watchObservedRunningTime="2026-03-13 16:01:01.486024436 +0000 UTC m=+6960.385812185" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.115842 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.116488 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.209069 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.546833 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:03 crc kubenswrapper[4907]: I0313 16:01:03.600925 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:01:04 crc kubenswrapper[4907]: E0313 16:01:04.040533 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode548a313_fc14_4ce9_b268_bf37c79eed83.slice/crio-4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979.scope\": RecentStats: unable to find data in memory cache]" Mar 13 16:01:04 crc kubenswrapper[4907]: I0313 16:01:04.508538 4907 generic.go:334] "Generic (PLEG): container finished" podID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerID="4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979" exitCode=0 Mar 13 16:01:04 crc kubenswrapper[4907]: I0313 16:01:04.508959 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerDied","Data":"4ff9398b9dec5963cf443cc668dffeb42de87e641ad4f61c8cd02708682d9979"} Mar 13 16:01:05 crc kubenswrapper[4907]: I0313 16:01:05.516214 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7nbgp" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" containerID="cri-o://5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" gracePeriod=2 Mar 13 16:01:05 crc kubenswrapper[4907]: I0313 16:01:05.893545 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.006709 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068337 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068516 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068710 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.068813 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") pod \"e548a313-fc14-4ce9-b268-bf37c79eed83\" (UID: \"e548a313-fc14-4ce9-b268-bf37c79eed83\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.073722 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.073786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz" (OuterVolumeSpecName: "kube-api-access-5qmmz") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "kube-api-access-5qmmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.096107 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.121636 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data" (OuterVolumeSpecName: "config-data") pod "e548a313-fc14-4ce9-b268-bf37c79eed83" (UID: "e548a313-fc14-4ce9-b268-bf37c79eed83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.170784 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") pod \"a13cade5-fe42-46a8-9c42-d1733bde2b78\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171089 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") pod \"a13cade5-fe42-46a8-9c42-d1733bde2b78\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171148 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") pod \"a13cade5-fe42-46a8-9c42-d1733bde2b78\" (UID: \"a13cade5-fe42-46a8-9c42-d1733bde2b78\") " Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171728 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qmmz\" (UniqueName: \"kubernetes.io/projected/e548a313-fc14-4ce9-b268-bf37c79eed83-kube-api-access-5qmmz\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171747 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171758 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171780 4907 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e548a313-fc14-4ce9-b268-bf37c79eed83-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.171968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities" (OuterVolumeSpecName: "utilities") pod "a13cade5-fe42-46a8-9c42-d1733bde2b78" (UID: "a13cade5-fe42-46a8-9c42-d1733bde2b78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.174476 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk" (OuterVolumeSpecName: "kube-api-access-x5stk") pod "a13cade5-fe42-46a8-9c42-d1733bde2b78" (UID: "a13cade5-fe42-46a8-9c42-d1733bde2b78"). InnerVolumeSpecName "kube-api-access-x5stk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.273508 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5stk\" (UniqueName: \"kubernetes.io/projected/a13cade5-fe42-46a8-9c42-d1733bde2b78-kube-api-access-x5stk\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.273543 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528013 4907 generic.go:334] "Generic (PLEG): container finished" podID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" exitCode=0 Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528055 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480"} Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528080 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nbgp" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528485 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nbgp" event={"ID":"a13cade5-fe42-46a8-9c42-d1733bde2b78","Type":"ContainerDied","Data":"673546a70a0ddd984900b7492ee18632405c8a55ae70039f43b9cdb5ae87218b"} Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.528515 4907 scope.go:117] "RemoveContainer" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.531156 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556961-gshbz" event={"ID":"e548a313-fc14-4ce9-b268-bf37c79eed83","Type":"ContainerDied","Data":"b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1"} Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.531192 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8e1101812a621dd17914318034354e00ada6a7f97ba9cc9819509f6432c55e1" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.531218 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556961-gshbz" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.553281 4907 scope.go:117] "RemoveContainer" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.587255 4907 scope.go:117] "RemoveContainer" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.612605 4907 scope.go:117] "RemoveContainer" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" Mar 13 16:01:06 crc kubenswrapper[4907]: E0313 16:01:06.613166 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480\": container with ID starting with 5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480 not found: ID does not exist" containerID="5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613306 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480"} err="failed to get container status \"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480\": rpc error: code = NotFound desc = could not find container \"5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480\": container with ID starting with 5770c77601b4306c58f1d10a7f3d071183506a543de6bfd4b22de746737a8480 not found: ID does not exist" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613388 4907 scope.go:117] "RemoveContainer" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" Mar 13 16:01:06 crc kubenswrapper[4907]: E0313 16:01:06.613768 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213\": container with ID starting with 526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213 not found: ID does not exist" containerID="526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613841 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213"} err="failed to get container status \"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213\": rpc error: code = NotFound desc = could not find container \"526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213\": container with ID starting with 526804910f6c587f26fe53ee4b1003b6d89146391dbb3182ce1f442e7cc49213 not found: ID does not exist" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.613922 4907 scope.go:117] "RemoveContainer" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" Mar 13 16:01:06 crc kubenswrapper[4907]: E0313 16:01:06.614465 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41\": container with ID starting with d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41 not found: ID does not exist" containerID="d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41" Mar 13 16:01:06 crc kubenswrapper[4907]: I0313 16:01:06.614512 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41"} err="failed to get container status \"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41\": rpc error: code = NotFound desc = could not find container \"d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41\": container with ID starting with d774a5d1c6b13e107977d37001f9f575455851b3923de08704f840ff5ea48c41 not found: ID does not exist" Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.234165 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a13cade5-fe42-46a8-9c42-d1733bde2b78" (UID: "a13cade5-fe42-46a8-9c42-d1733bde2b78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.295275 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a13cade5-fe42-46a8-9c42-d1733bde2b78-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.474944 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.491132 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7nbgp"] Mar 13 16:01:07 crc kubenswrapper[4907]: I0313 16:01:07.797377 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" path="/var/lib/kubelet/pods/a13cade5-fe42-46a8-9c42-d1733bde2b78/volumes" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.041790 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.042349 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.042408 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.043299 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.043351 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19" gracePeriod=600 Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.674862 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19" exitCode=0 Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.675223 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19"} Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.675251 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964"} Mar 13 16:01:18 crc kubenswrapper[4907]: I0313 16:01:18.675268 4907 scope.go:117] "RemoveContainer" containerID="33f8e315523143b157797243610ad7b82c7815f9e551d1169f2c036cfcadf792" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.143833 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144791 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144806 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144827 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-content" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144833 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-content" Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144859 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-utilities" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144867 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="extract-utilities" Mar 13 16:02:00 crc kubenswrapper[4907]: E0313 16:02:00.144917 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerName="keystone-cron" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.144926 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerName="keystone-cron" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.145140 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a13cade5-fe42-46a8-9c42-d1733bde2b78" containerName="registry-server" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.145170 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e548a313-fc14-4ce9-b268-bf37c79eed83" containerName="keystone-cron" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.146014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.148912 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.149100 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.149369 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.155196 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.215408 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"auto-csr-approver-29556962-d4bk5\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.318080 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"auto-csr-approver-29556962-d4bk5\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.336353 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"auto-csr-approver-29556962-d4bk5\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.466979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:00 crc kubenswrapper[4907]: I0313 16:02:00.938286 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:02:01 crc kubenswrapper[4907]: I0313 16:02:01.047290 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" event={"ID":"927835ce-92e1-4a29-b064-0e08a5288442","Type":"ContainerStarted","Data":"59176c1fbdb0435bb688c727d47f358b21ee89c1dba9385298bbb961b4960ce1"} Mar 13 16:02:03 crc kubenswrapper[4907]: I0313 16:02:03.066648 4907 generic.go:334] "Generic (PLEG): container finished" podID="927835ce-92e1-4a29-b064-0e08a5288442" containerID="fd8d416a2c11ed375f8397b31e95379985b54fa1a07f700d79a805d7106a6c4e" exitCode=0 Mar 13 16:02:03 crc kubenswrapper[4907]: I0313 16:02:03.066767 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" event={"ID":"927835ce-92e1-4a29-b064-0e08a5288442","Type":"ContainerDied","Data":"fd8d416a2c11ed375f8397b31e95379985b54fa1a07f700d79a805d7106a6c4e"} Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.452952 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.623664 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") pod \"927835ce-92e1-4a29-b064-0e08a5288442\" (UID: \"927835ce-92e1-4a29-b064-0e08a5288442\") " Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.629233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8" (OuterVolumeSpecName: "kube-api-access-7wgf8") pod "927835ce-92e1-4a29-b064-0e08a5288442" (UID: "927835ce-92e1-4a29-b064-0e08a5288442"). InnerVolumeSpecName "kube-api-access-7wgf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:02:04 crc kubenswrapper[4907]: I0313 16:02:04.726707 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wgf8\" (UniqueName: \"kubernetes.io/projected/927835ce-92e1-4a29-b064-0e08a5288442-kube-api-access-7wgf8\") on node \"crc\" DevicePath \"\"" Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.084539 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" event={"ID":"927835ce-92e1-4a29-b064-0e08a5288442","Type":"ContainerDied","Data":"59176c1fbdb0435bb688c727d47f358b21ee89c1dba9385298bbb961b4960ce1"} Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.084578 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59176c1fbdb0435bb688c727d47f358b21ee89c1dba9385298bbb961b4960ce1" Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.084638 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556962-d4bk5" Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.527560 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.537656 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556956-txd4m"] Mar 13 16:02:05 crc kubenswrapper[4907]: I0313 16:02:05.797256 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a22eedb1-70a8-4232-9a54-8625a4dd7022" path="/var/lib/kubelet/pods/a22eedb1-70a8-4232-9a54-8625a4dd7022/volumes" Mar 13 16:02:40 crc kubenswrapper[4907]: I0313 16:02:40.789789 4907 scope.go:117] "RemoveContainer" containerID="76fd8f67a17de0068ab3b4548c472b268fc93309cdb749646d71dbe2fb8aef2e" Mar 13 16:03:18 crc kubenswrapper[4907]: I0313 16:03:18.042496 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:03:18 crc kubenswrapper[4907]: I0313 16:03:18.043414 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:03:29 crc kubenswrapper[4907]: I0313 16:03:29.038535 4907 generic.go:334] "Generic (PLEG): container finished" podID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerID="384d76deb615542ee9afab32cbd864c5a2ef0e524a3354232d127d85ac034766" exitCode=0 Mar 13 16:03:29 crc kubenswrapper[4907]: I0313 16:03:29.038646 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerDied","Data":"384d76deb615542ee9afab32cbd864c5a2ef0e524a3354232d127d85ac034766"} Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.708986 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.830688 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831252 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831440 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831570 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.831674 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") pod \"02ef8f0f-4ae4-478a-a729-3bde4adade43\" (UID: \"02ef8f0f-4ae4-478a-a729-3bde4adade43\") " Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.837256 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph" (OuterVolumeSpecName: "ceph") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.837611 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.838740 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk" (OuterVolumeSpecName: "kube-api-access-gsfkk") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "kube-api-access-gsfkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.872304 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory" (OuterVolumeSpecName: "inventory") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.872394 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "02ef8f0f-4ae4-478a-a729-3bde4adade43" (UID: "02ef8f0f-4ae4-478a-a729-3bde4adade43"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935154 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935199 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsfkk\" (UniqueName: \"kubernetes.io/projected/02ef8f0f-4ae4-478a-a729-3bde4adade43-kube-api-access-gsfkk\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935220 4907 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935239 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:30 crc kubenswrapper[4907]: I0313 16:03:30.935261 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02ef8f0f-4ae4-478a-a729-3bde4adade43-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:31 crc kubenswrapper[4907]: I0313 16:03:31.061698 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" event={"ID":"02ef8f0f-4ae4-478a-a729-3bde4adade43","Type":"ContainerDied","Data":"4e306d67a326c5a22924c14c80f0301c152e3eb9bef16aa1cc486b8e33a33cc7"} Mar 13 16:03:31 crc kubenswrapper[4907]: I0313 16:03:31.061947 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e306d67a326c5a22924c14c80f0301c152e3eb9bef16aa1cc486b8e33a33cc7" Mar 13 16:03:31 crc kubenswrapper[4907]: I0313 16:03:31.061728 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.712957 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-tb9sw"] Mar 13 16:03:36 crc kubenswrapper[4907]: E0313 16:03:36.713840 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.713853 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: E0313 16:03:36.713869 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="927835ce-92e1-4a29-b064-0e08a5288442" containerName="oc" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.713889 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="927835ce-92e1-4a29-b064-0e08a5288442" containerName="oc" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.714165 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="02ef8f0f-4ae4-478a-a729-3bde4adade43" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.714183 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="927835ce-92e1-4a29-b064-0e08a5288442" containerName="oc" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.714926 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.719779 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.719796 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.721313 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.721345 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.732379 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-tb9sw"] Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.870120 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.871050 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.871118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.872118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.872404 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974259 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974389 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974433 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974471 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.974497 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.980710 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.981384 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.982342 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:36 crc kubenswrapper[4907]: I0313 16:03:36.982554 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:37 crc kubenswrapper[4907]: I0313 16:03:37.002779 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"bootstrap-openstack-openstack-cell1-tb9sw\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:37 crc kubenswrapper[4907]: I0313 16:03:37.035677 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:03:37 crc kubenswrapper[4907]: I0313 16:03:37.605183 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-tb9sw"] Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.166074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerStarted","Data":"239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777"} Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.596151 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.598791 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.608475 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.706992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.707061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.707275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.809249 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.809322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.809384 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.810084 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.810340 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.835166 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"redhat-marketplace-frdhh\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:38 crc kubenswrapper[4907]: I0313 16:03:38.923035 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:39 crc kubenswrapper[4907]: I0313 16:03:39.176394 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerStarted","Data":"709051d88f8ea62f2c110585bd6b47f69dfdd1063e53bf548cb2db80a34cddd9"} Mar 13 16:03:39 crc kubenswrapper[4907]: I0313 16:03:39.199042 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" podStartSLOduration=2.799800523 podStartE2EDuration="3.199024336s" podCreationTimestamp="2026-03-13 16:03:36 +0000 UTC" firstStartedPulling="2026-03-13 16:03:37.616083567 +0000 UTC m=+7116.515871256" lastFinishedPulling="2026-03-13 16:03:38.01530736 +0000 UTC m=+7116.915095069" observedRunningTime="2026-03-13 16:03:39.193914256 +0000 UTC m=+7118.093701975" watchObservedRunningTime="2026-03-13 16:03:39.199024336 +0000 UTC m=+7118.098812015" Mar 13 16:03:39 crc kubenswrapper[4907]: I0313 16:03:39.411019 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:40 crc kubenswrapper[4907]: I0313 16:03:40.188762 4907 generic.go:334] "Generic (PLEG): container finished" podID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerID="fa32b7f26cf0790f4f10f343dafcb6fe343d4f8cfbb9a8065a1c17955425e202" exitCode=0 Mar 13 16:03:40 crc kubenswrapper[4907]: I0313 16:03:40.189697 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"fa32b7f26cf0790f4f10f343dafcb6fe343d4f8cfbb9a8065a1c17955425e202"} Mar 13 16:03:40 crc kubenswrapper[4907]: I0313 16:03:40.189759 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerStarted","Data":"f1c6800484e1fbe1a1b02926ef17c56558da8199ea2caad3fd5b7d3de1859a09"} Mar 13 16:03:43 crc kubenswrapper[4907]: I0313 16:03:43.238930 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerStarted","Data":"ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77"} Mar 13 16:03:44 crc kubenswrapper[4907]: I0313 16:03:44.248926 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77"} Mar 13 16:03:44 crc kubenswrapper[4907]: I0313 16:03:44.248868 4907 generic.go:334] "Generic (PLEG): container finished" podID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerID="ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77" exitCode=0 Mar 13 16:03:45 crc kubenswrapper[4907]: I0313 16:03:45.261459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerStarted","Data":"f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c"} Mar 13 16:03:45 crc kubenswrapper[4907]: I0313 16:03:45.283190 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-frdhh" podStartSLOduration=2.7292045910000002 podStartE2EDuration="7.283170972s" podCreationTimestamp="2026-03-13 16:03:38 +0000 UTC" firstStartedPulling="2026-03-13 16:03:40.19131186 +0000 UTC m=+7119.091099549" lastFinishedPulling="2026-03-13 16:03:44.745278241 +0000 UTC m=+7123.645065930" observedRunningTime="2026-03-13 16:03:45.278893765 +0000 UTC m=+7124.178681474" watchObservedRunningTime="2026-03-13 16:03:45.283170972 +0000 UTC m=+7124.182958671" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.042401 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.044995 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.924485 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.924797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:48 crc kubenswrapper[4907]: I0313 16:03:48.973340 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:49 crc kubenswrapper[4907]: I0313 16:03:49.348651 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.063731 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.064470 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-frdhh" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" containerID="cri-o://f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c" gracePeriod=2 Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.364640 4907 generic.go:334] "Generic (PLEG): container finished" podID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerID="f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c" exitCode=0 Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.364730 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c"} Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.527502 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.603518 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") pod \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.603779 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") pod \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.603939 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") pod \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\" (UID: \"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc\") " Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.604818 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities" (OuterVolumeSpecName: "utilities") pod "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" (UID: "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.614135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf" (OuterVolumeSpecName: "kube-api-access-g57nf") pod "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" (UID: "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc"). InnerVolumeSpecName "kube-api-access-g57nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.650966 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" (UID: "8371fb8e-84f3-4e9a-a810-9ab7ed345cbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.707272 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g57nf\" (UniqueName: \"kubernetes.io/projected/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-kube-api-access-g57nf\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.707317 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:52 crc kubenswrapper[4907]: I0313 16:03:52.707330 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.377478 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-frdhh" event={"ID":"8371fb8e-84f3-4e9a-a810-9ab7ed345cbc","Type":"ContainerDied","Data":"f1c6800484e1fbe1a1b02926ef17c56558da8199ea2caad3fd5b7d3de1859a09"} Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.377911 4907 scope.go:117] "RemoveContainer" containerID="f0fb1ab8a39a940de2067640271d97e3a6053bce6138d1bb4cd8baad8354c46c" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.377527 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-frdhh" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.414322 4907 scope.go:117] "RemoveContainer" containerID="ec8675a76aec53278dabecd965f5899500f4f28f9bc79e544593e636a482ce77" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.433735 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.441400 4907 scope.go:117] "RemoveContainer" containerID="fa32b7f26cf0790f4f10f343dafcb6fe343d4f8cfbb9a8065a1c17955425e202" Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.447184 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-frdhh"] Mar 13 16:03:53 crc kubenswrapper[4907]: I0313 16:03:53.797340 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" path="/var/lib/kubelet/pods/8371fb8e-84f3-4e9a-a810-9ab7ed345cbc/volumes" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.163954 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:04:00 crc kubenswrapper[4907]: E0313 16:04:00.165785 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.165822 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" Mar 13 16:04:00 crc kubenswrapper[4907]: E0313 16:04:00.165850 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-utilities" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.165867 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-utilities" Mar 13 16:04:00 crc kubenswrapper[4907]: E0313 16:04:00.165946 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-content" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.165968 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="extract-content" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.166472 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8371fb8e-84f3-4e9a-a810-9ab7ed345cbc" containerName="registry-server" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.167916 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.172334 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.172492 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.172566 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.181611 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.205874 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"auto-csr-approver-29556964-gfm8d\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.308369 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"auto-csr-approver-29556964-gfm8d\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.327208 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"auto-csr-approver-29556964-gfm8d\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:00 crc kubenswrapper[4907]: I0313 16:04:00.506475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:01 crc kubenswrapper[4907]: I0313 16:04:01.046074 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:04:01 crc kubenswrapper[4907]: W0313 16:04:01.057132 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod969a7b07_f302_48b0_b48e_730836ea94a4.slice/crio-997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7 WatchSource:0}: Error finding container 997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7: Status 404 returned error can't find the container with id 997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7 Mar 13 16:04:01 crc kubenswrapper[4907]: I0313 16:04:01.480228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" event={"ID":"969a7b07-f302-48b0-b48e-730836ea94a4","Type":"ContainerStarted","Data":"997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7"} Mar 13 16:04:03 crc kubenswrapper[4907]: I0313 16:04:03.506293 4907 generic.go:334] "Generic (PLEG): container finished" podID="969a7b07-f302-48b0-b48e-730836ea94a4" containerID="e87bfe7ae00d8363f28ac3bbec93999ba43f3034b744a941e08c8397ba4ba459" exitCode=0 Mar 13 16:04:03 crc kubenswrapper[4907]: I0313 16:04:03.506357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" event={"ID":"969a7b07-f302-48b0-b48e-730836ea94a4","Type":"ContainerDied","Data":"e87bfe7ae00d8363f28ac3bbec93999ba43f3034b744a941e08c8397ba4ba459"} Mar 13 16:04:04 crc kubenswrapper[4907]: I0313 16:04:04.918925 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.026833 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") pod \"969a7b07-f302-48b0-b48e-730836ea94a4\" (UID: \"969a7b07-f302-48b0-b48e-730836ea94a4\") " Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.033684 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p" (OuterVolumeSpecName: "kube-api-access-lqn6p") pod "969a7b07-f302-48b0-b48e-730836ea94a4" (UID: "969a7b07-f302-48b0-b48e-730836ea94a4"). InnerVolumeSpecName "kube-api-access-lqn6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.129435 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqn6p\" (UniqueName: \"kubernetes.io/projected/969a7b07-f302-48b0-b48e-730836ea94a4-kube-api-access-lqn6p\") on node \"crc\" DevicePath \"\"" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.531179 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" event={"ID":"969a7b07-f302-48b0-b48e-730836ea94a4","Type":"ContainerDied","Data":"997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7"} Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.531255 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="997b3fe277f17cc809d2ed0464d2c10cc99ea1e615bd392ec2c3ba14ec9dd0f7" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.531304 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556964-gfm8d" Mar 13 16:04:05 crc kubenswrapper[4907]: I0313 16:04:05.992205 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 16:04:06 crc kubenswrapper[4907]: I0313 16:04:06.001259 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556958-dk8dz"] Mar 13 16:04:07 crc kubenswrapper[4907]: I0313 16:04:07.793948 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d38cd8e-5a08-4c56-b35e-88617df1d3a8" path="/var/lib/kubelet/pods/2d38cd8e-5a08-4c56-b35e-88617df1d3a8/volumes" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.041186 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.041823 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.041911 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.042724 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.042844 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" gracePeriod=600 Mar 13 16:04:18 crc kubenswrapper[4907]: E0313 16:04:18.173978 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.657452 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" exitCode=0 Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.657580 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964"} Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.658135 4907 scope.go:117] "RemoveContainer" containerID="2a3c556bf9f1ef9e4af9dec2c5d043c5ea60bb781b3c04fb7c2299179fb58a19" Mar 13 16:04:18 crc kubenswrapper[4907]: I0313 16:04:18.659014 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:18 crc kubenswrapper[4907]: E0313 16:04:18.659678 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:29 crc kubenswrapper[4907]: I0313 16:04:29.782809 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:29 crc kubenswrapper[4907]: E0313 16:04:29.783790 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:40 crc kubenswrapper[4907]: I0313 16:04:40.901229 4907 scope.go:117] "RemoveContainer" containerID="d5d998084a23eab3592589afda6539d62cf5a3559ead1f830c6d581d1f9410c7" Mar 13 16:04:42 crc kubenswrapper[4907]: I0313 16:04:42.782577 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:42 crc kubenswrapper[4907]: E0313 16:04:42.784149 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:04:57 crc kubenswrapper[4907]: I0313 16:04:57.782348 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:04:57 crc kubenswrapper[4907]: E0313 16:04:57.783161 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:10 crc kubenswrapper[4907]: I0313 16:05:10.782165 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:10 crc kubenswrapper[4907]: E0313 16:05:10.783216 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:23 crc kubenswrapper[4907]: I0313 16:05:23.785366 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:23 crc kubenswrapper[4907]: E0313 16:05:23.786255 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:38 crc kubenswrapper[4907]: I0313 16:05:38.782507 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:38 crc kubenswrapper[4907]: E0313 16:05:38.783502 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:05:53 crc kubenswrapper[4907]: I0313 16:05:53.783766 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:05:53 crc kubenswrapper[4907]: E0313 16:05:53.784712 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.164764 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:06:00 crc kubenswrapper[4907]: E0313 16:06:00.166214 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" containerName="oc" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.166237 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" containerName="oc" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.166701 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" containerName="oc" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.168256 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.170671 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.171081 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.171098 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.178345 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.199362 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"auto-csr-approver-29556966-drz8k\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.301793 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"auto-csr-approver-29556966-drz8k\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.321411 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"auto-csr-approver-29556966-drz8k\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.496345 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:00 crc kubenswrapper[4907]: W0313 16:06:00.990363 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e6bd258_54f3_49e8_8b92_d101e6b19872.slice/crio-a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93 WatchSource:0}: Error finding container a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93: Status 404 returned error can't find the container with id a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93 Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.990534 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:06:00 crc kubenswrapper[4907]: I0313 16:06:00.994008 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:06:01 crc kubenswrapper[4907]: I0313 16:06:01.847757 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556966-drz8k" event={"ID":"9e6bd258-54f3-49e8-8b92-d101e6b19872","Type":"ContainerStarted","Data":"a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93"} Mar 13 16:06:02 crc kubenswrapper[4907]: I0313 16:06:02.860700 4907 generic.go:334] "Generic (PLEG): container finished" podID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerID="fba4ffaa28135dbeabe5c056323b0e99fffec19d264cb5778eaf0cb6c3cc2d40" exitCode=0 Mar 13 16:06:02 crc kubenswrapper[4907]: I0313 16:06:02.860788 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556966-drz8k" event={"ID":"9e6bd258-54f3-49e8-8b92-d101e6b19872","Type":"ContainerDied","Data":"fba4ffaa28135dbeabe5c056323b0e99fffec19d264cb5778eaf0cb6c3cc2d40"} Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.347316 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.387159 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") pod \"9e6bd258-54f3-49e8-8b92-d101e6b19872\" (UID: \"9e6bd258-54f3-49e8-8b92-d101e6b19872\") " Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.395138 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm" (OuterVolumeSpecName: "kube-api-access-ttfcm") pod "9e6bd258-54f3-49e8-8b92-d101e6b19872" (UID: "9e6bd258-54f3-49e8-8b92-d101e6b19872"). InnerVolumeSpecName "kube-api-access-ttfcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.489442 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttfcm\" (UniqueName: \"kubernetes.io/projected/9e6bd258-54f3-49e8-8b92-d101e6b19872-kube-api-access-ttfcm\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.911006 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556966-drz8k" event={"ID":"9e6bd258-54f3-49e8-8b92-d101e6b19872","Type":"ContainerDied","Data":"a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93"} Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.911077 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a293a2a950339887882ecd7a9d1f5edab077c3fee5d7b369f1ab1d20a5402f93" Mar 13 16:06:04 crc kubenswrapper[4907]: I0313 16:06:04.911178 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556966-drz8k" Mar 13 16:06:05 crc kubenswrapper[4907]: I0313 16:06:05.447327 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:06:05 crc kubenswrapper[4907]: I0313 16:06:05.455324 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556960-8xc84"] Mar 13 16:06:05 crc kubenswrapper[4907]: I0313 16:06:05.797677 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acb3140e-a115-443a-a39e-5f180a196b13" path="/var/lib/kubelet/pods/acb3140e-a115-443a-a39e-5f180a196b13/volumes" Mar 13 16:06:08 crc kubenswrapper[4907]: I0313 16:06:08.782508 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:08 crc kubenswrapper[4907]: E0313 16:06:08.783268 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:22 crc kubenswrapper[4907]: I0313 16:06:22.782172 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:22 crc kubenswrapper[4907]: E0313 16:06:22.783015 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:37 crc kubenswrapper[4907]: I0313 16:06:37.782553 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:37 crc kubenswrapper[4907]: E0313 16:06:37.783581 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:41 crc kubenswrapper[4907]: I0313 16:06:41.030460 4907 scope.go:117] "RemoveContainer" containerID="90c713255623c6a0cd4f0904c875e6fe6995c5ac18245db4d00a77bd7e6d5d8c" Mar 13 16:06:51 crc kubenswrapper[4907]: I0313 16:06:51.377567 4907 generic.go:334] "Generic (PLEG): container finished" podID="d7346ccb-6154-42a9-9097-bad966681d81" containerID="709051d88f8ea62f2c110585bd6b47f69dfdd1063e53bf548cb2db80a34cddd9" exitCode=0 Mar 13 16:06:51 crc kubenswrapper[4907]: I0313 16:06:51.377700 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerDied","Data":"709051d88f8ea62f2c110585bd6b47f69dfdd1063e53bf548cb2db80a34cddd9"} Mar 13 16:06:51 crc kubenswrapper[4907]: I0313 16:06:51.788336 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:06:51 crc kubenswrapper[4907]: E0313 16:06:51.789817 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.879459 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.990541 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.990611 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.990922 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.991043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.991175 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") pod \"d7346ccb-6154-42a9-9097-bad966681d81\" (UID: \"d7346ccb-6154-42a9-9097-bad966681d81\") " Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.996500 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:52 crc kubenswrapper[4907]: I0313 16:06:52.998047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph" (OuterVolumeSpecName: "ceph") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.001255 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb" (OuterVolumeSpecName: "kube-api-access-8bpfb") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "kube-api-access-8bpfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.023287 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory" (OuterVolumeSpecName: "inventory") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.029706 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "d7346ccb-6154-42a9-9097-bad966681d81" (UID: "d7346ccb-6154-42a9-9097-bad966681d81"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097463 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097500 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097512 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097522 4907 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7346ccb-6154-42a9-9097-bad966681d81-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.097532 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bpfb\" (UniqueName: \"kubernetes.io/projected/d7346ccb-6154-42a9-9097-bad966681d81-kube-api-access-8bpfb\") on node \"crc\" DevicePath \"\"" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.404013 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" event={"ID":"d7346ccb-6154-42a9-9097-bad966681d81","Type":"ContainerDied","Data":"239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777"} Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.404060 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-tb9sw" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.404081 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.543967 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-84zm8"] Mar 13 16:06:53 crc kubenswrapper[4907]: E0313 16:06:53.544502 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerName="oc" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544523 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerName="oc" Mar 13 16:06:53 crc kubenswrapper[4907]: E0313 16:06:53.544558 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7346ccb-6154-42a9-9097-bad966681d81" containerName="bootstrap-openstack-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544566 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7346ccb-6154-42a9-9097-bad966681d81" containerName="bootstrap-openstack-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544842 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7346ccb-6154-42a9-9097-bad966681d81" containerName="bootstrap-openstack-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.544897 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" containerName="oc" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.545789 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.551499 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.551810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.552041 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.552493 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.574948 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-84zm8"] Mar 13 16:06:53 crc kubenswrapper[4907]: E0313 16:06:53.580211 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7346ccb_6154_42a9_9097_bad966681d81.slice/crio-239301abbc130c5a894d3a44690a47910fda2c931b2896b41f7ec97e0db3d777\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7346ccb_6154_42a9_9097_bad966681d81.slice\": RecentStats: unable to find data in memory cache]" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.616680 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.616786 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.616984 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.617086 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719245 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719334 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719398 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.719450 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.732668 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.732964 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.739462 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.755941 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"download-cache-openstack-openstack-cell1-84zm8\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:53 crc kubenswrapper[4907]: I0313 16:06:53.870124 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:06:54 crc kubenswrapper[4907]: I0313 16:06:54.437077 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-84zm8"] Mar 13 16:06:55 crc kubenswrapper[4907]: I0313 16:06:55.422614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerStarted","Data":"1047ba81465e9116e35201b4b46e027b0e8aa7f144dbbe56c1ec6cbfadc3afa4"} Mar 13 16:06:55 crc kubenswrapper[4907]: I0313 16:06:55.423058 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerStarted","Data":"a2c0ec2fb6e0a54b73a9abdf461f21aec42d36ae856307b5fc9e10ec95882a6b"} Mar 13 16:06:55 crc kubenswrapper[4907]: I0313 16:06:55.454997 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" podStartSLOduration=1.9359263119999999 podStartE2EDuration="2.454978757s" podCreationTimestamp="2026-03-13 16:06:53 +0000 UTC" firstStartedPulling="2026-03-13 16:06:54.447318523 +0000 UTC m=+7313.347106212" lastFinishedPulling="2026-03-13 16:06:54.966370958 +0000 UTC m=+7313.866158657" observedRunningTime="2026-03-13 16:06:55.448566022 +0000 UTC m=+7314.348353741" watchObservedRunningTime="2026-03-13 16:06:55.454978757 +0000 UTC m=+7314.354766446" Mar 13 16:07:06 crc kubenswrapper[4907]: I0313 16:07:06.782748 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:06 crc kubenswrapper[4907]: E0313 16:07:06.783602 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:19 crc kubenswrapper[4907]: I0313 16:07:19.783489 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:19 crc kubenswrapper[4907]: E0313 16:07:19.784235 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:33 crc kubenswrapper[4907]: I0313 16:07:33.783402 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:33 crc kubenswrapper[4907]: E0313 16:07:33.784308 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:44 crc kubenswrapper[4907]: I0313 16:07:44.782058 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:44 crc kubenswrapper[4907]: E0313 16:07:44.782860 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:07:57 crc kubenswrapper[4907]: I0313 16:07:57.784103 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:07:57 crc kubenswrapper[4907]: E0313 16:07:57.785235 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.176491 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.179513 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.183673 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.184421 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.185359 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.195104 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.260496 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"auto-csr-approver-29556968-xbmhk\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.363105 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"auto-csr-approver-29556968-xbmhk\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.393204 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"auto-csr-approver-29556968-xbmhk\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:00 crc kubenswrapper[4907]: I0313 16:08:00.514284 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:01 crc kubenswrapper[4907]: I0313 16:08:01.010492 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:08:01 crc kubenswrapper[4907]: W0313 16:08:01.018786 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfdfc46b_b215_491c_ab2f_2a108c2c0882.slice/crio-2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490 WatchSource:0}: Error finding container 2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490: Status 404 returned error can't find the container with id 2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490 Mar 13 16:08:01 crc kubenswrapper[4907]: I0313 16:08:01.125300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" event={"ID":"cfdfc46b-b215-491c-ab2f-2a108c2c0882","Type":"ContainerStarted","Data":"2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490"} Mar 13 16:08:03 crc kubenswrapper[4907]: I0313 16:08:03.152090 4907 generic.go:334] "Generic (PLEG): container finished" podID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerID="6f4aac0052e053a0aefd7363aac839de6a2ba5c8481dbf090c19d76f2294e95d" exitCode=0 Mar 13 16:08:03 crc kubenswrapper[4907]: I0313 16:08:03.152245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" event={"ID":"cfdfc46b-b215-491c-ab2f-2a108c2c0882","Type":"ContainerDied","Data":"6f4aac0052e053a0aefd7363aac839de6a2ba5c8481dbf090c19d76f2294e95d"} Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.627774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.681313 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") pod \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\" (UID: \"cfdfc46b-b215-491c-ab2f-2a108c2c0882\") " Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.687991 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f" (OuterVolumeSpecName: "kube-api-access-vks5f") pod "cfdfc46b-b215-491c-ab2f-2a108c2c0882" (UID: "cfdfc46b-b215-491c-ab2f-2a108c2c0882"). InnerVolumeSpecName "kube-api-access-vks5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:08:04 crc kubenswrapper[4907]: I0313 16:08:04.783967 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vks5f\" (UniqueName: \"kubernetes.io/projected/cfdfc46b-b215-491c-ab2f-2a108c2c0882-kube-api-access-vks5f\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.176005 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" event={"ID":"cfdfc46b-b215-491c-ab2f-2a108c2c0882","Type":"ContainerDied","Data":"2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490"} Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.176426 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c4008c2c653fef7d8f4f2a44b0be2fffd84b61991c523324d0a0429b4aa4490" Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.176049 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556968-xbmhk" Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.704428 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.716412 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556962-d4bk5"] Mar 13 16:08:05 crc kubenswrapper[4907]: I0313 16:08:05.814594 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="927835ce-92e1-4a29-b064-0e08a5288442" path="/var/lib/kubelet/pods/927835ce-92e1-4a29-b064-0e08a5288442/volumes" Mar 13 16:08:11 crc kubenswrapper[4907]: I0313 16:08:11.800651 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:11 crc kubenswrapper[4907]: E0313 16:08:11.801957 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:24 crc kubenswrapper[4907]: I0313 16:08:24.782306 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:24 crc kubenswrapper[4907]: E0313 16:08:24.783222 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:36 crc kubenswrapper[4907]: I0313 16:08:36.783526 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:36 crc kubenswrapper[4907]: E0313 16:08:36.784287 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:41 crc kubenswrapper[4907]: I0313 16:08:41.130166 4907 scope.go:117] "RemoveContainer" containerID="fd8d416a2c11ed375f8397b31e95379985b54fa1a07f700d79a805d7106a6c4e" Mar 13 16:08:50 crc kubenswrapper[4907]: I0313 16:08:50.784173 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:08:50 crc kubenswrapper[4907]: E0313 16:08:50.784912 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:08:54 crc kubenswrapper[4907]: I0313 16:08:54.882833 4907 generic.go:334] "Generic (PLEG): container finished" podID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerID="1047ba81465e9116e35201b4b46e027b0e8aa7f144dbbe56c1ec6cbfadc3afa4" exitCode=0 Mar 13 16:08:54 crc kubenswrapper[4907]: I0313 16:08:54.882917 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerDied","Data":"1047ba81465e9116e35201b4b46e027b0e8aa7f144dbbe56c1ec6cbfadc3afa4"} Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.323703 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.497654 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.497832 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.498018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.498204 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") pod \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\" (UID: \"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9\") " Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.506559 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph" (OuterVolumeSpecName: "ceph") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.506747 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln" (OuterVolumeSpecName: "kube-api-access-2jvln") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "kube-api-access-2jvln". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.527294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.532972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory" (OuterVolumeSpecName: "inventory") pod "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" (UID: "441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600824 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600855 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600867 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jvln\" (UniqueName: \"kubernetes.io/projected/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-kube-api-access-2jvln\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.600875 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.901762 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" event={"ID":"441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9","Type":"ContainerDied","Data":"a2c0ec2fb6e0a54b73a9abdf461f21aec42d36ae856307b5fc9e10ec95882a6b"} Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.901799 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c0ec2fb6e0a54b73a9abdf461f21aec42d36ae856307b5fc9e10ec95882a6b" Mar 13 16:08:56 crc kubenswrapper[4907]: I0313 16:08:56.901806 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-84zm8" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.000440 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-wrvtc"] Mar 13 16:08:57 crc kubenswrapper[4907]: E0313 16:08:57.000913 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerName="download-cache-openstack-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.000929 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerName="download-cache-openstack-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: E0313 16:08:57.000955 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerName="oc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.000961 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerName="oc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.001164 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9" containerName="download-cache-openstack-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.001201 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" containerName="oc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.001942 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.006011 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.006064 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.006730 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.008202 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.019278 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-wrvtc"] Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.111466 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.111794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.111903 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.112032 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214397 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214548 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214669 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.214698 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.220722 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.220938 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.226012 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.232533 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"configure-network-openstack-openstack-cell1-wrvtc\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.373597 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:08:57 crc kubenswrapper[4907]: I0313 16:08:57.924736 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-wrvtc"] Mar 13 16:08:58 crc kubenswrapper[4907]: I0313 16:08:58.925441 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerStarted","Data":"af77d9f29bb5e409e89eed0618c3878a4fc70669c940d590206ec36dfdf086f6"} Mar 13 16:08:58 crc kubenswrapper[4907]: I0313 16:08:58.925857 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerStarted","Data":"ff5e634a1d2fde12b87f2739aa06c6e49877afb509500e105fbf347bc7380990"} Mar 13 16:08:58 crc kubenswrapper[4907]: I0313 16:08:58.953381 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" podStartSLOduration=2.46910909 podStartE2EDuration="2.953362101s" podCreationTimestamp="2026-03-13 16:08:56 +0000 UTC" firstStartedPulling="2026-03-13 16:08:57.936360451 +0000 UTC m=+7436.836148150" lastFinishedPulling="2026-03-13 16:08:58.420613472 +0000 UTC m=+7437.320401161" observedRunningTime="2026-03-13 16:08:58.94744873 +0000 UTC m=+7437.847236509" watchObservedRunningTime="2026-03-13 16:08:58.953362101 +0000 UTC m=+7437.853149790" Mar 13 16:09:02 crc kubenswrapper[4907]: I0313 16:09:02.783311 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:09:02 crc kubenswrapper[4907]: E0313 16:09:02.784576 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:09:15 crc kubenswrapper[4907]: I0313 16:09:15.783484 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:09:15 crc kubenswrapper[4907]: E0313 16:09:15.785987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:09:29 crc kubenswrapper[4907]: I0313 16:09:29.783786 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:09:30 crc kubenswrapper[4907]: I0313 16:09:30.270012 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560"} Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.238818 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.246547 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.253432 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.447794 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.448029 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.448092 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.552461 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.552873 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.553317 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.555926 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.556406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.587698 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"redhat-operators-4vfxw\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:54 crc kubenswrapper[4907]: I0313 16:09:54.595271 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.085181 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.567652 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" exitCode=0 Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.567704 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5"} Mar 13 16:09:55 crc kubenswrapper[4907]: I0313 16:09:55.567965 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerStarted","Data":"c899082ad7772f673db6c6ad02b7f849f4e3f906cbceceed2c5b27f98250c37d"} Mar 13 16:09:57 crc kubenswrapper[4907]: I0313 16:09:57.590592 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerStarted","Data":"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60"} Mar 13 16:09:58 crc kubenswrapper[4907]: I0313 16:09:58.598974 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" exitCode=0 Mar 13 16:09:58 crc kubenswrapper[4907]: I0313 16:09:58.600120 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60"} Mar 13 16:09:59 crc kubenswrapper[4907]: I0313 16:09:59.614354 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerStarted","Data":"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8"} Mar 13 16:09:59 crc kubenswrapper[4907]: I0313 16:09:59.645406 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4vfxw" podStartSLOduration=2.182975002 podStartE2EDuration="5.64538447s" podCreationTimestamp="2026-03-13 16:09:54 +0000 UTC" firstStartedPulling="2026-03-13 16:09:55.571195908 +0000 UTC m=+7494.470983587" lastFinishedPulling="2026-03-13 16:09:59.033605366 +0000 UTC m=+7497.933393055" observedRunningTime="2026-03-13 16:09:59.636365203 +0000 UTC m=+7498.536152962" watchObservedRunningTime="2026-03-13 16:09:59.64538447 +0000 UTC m=+7498.545172169" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.148039 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.149560 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.152567 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.153040 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.154322 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.157963 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.196484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"auto-csr-approver-29556970-lbxp9\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.297914 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"auto-csr-approver-29556970-lbxp9\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.317623 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"auto-csr-approver-29556970-lbxp9\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.479922 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:00 crc kubenswrapper[4907]: I0313 16:10:00.981326 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:10:01 crc kubenswrapper[4907]: I0313 16:10:01.673852 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" event={"ID":"2b3a466a-0c1c-4c86-897f-0a3375a9fb53","Type":"ContainerStarted","Data":"6da23a5e86abc141883193c299ad8e19cb1959fa6ac9e21e18427d7c1e087244"} Mar 13 16:10:03 crc kubenswrapper[4907]: I0313 16:10:03.701584 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerID="1ffb094832567d23bf593fb1872edfb355e12059011cdaa6ddadd04821d723cc" exitCode=0 Mar 13 16:10:03 crc kubenswrapper[4907]: I0313 16:10:03.701667 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" event={"ID":"2b3a466a-0c1c-4c86-897f-0a3375a9fb53","Type":"ContainerDied","Data":"1ffb094832567d23bf593fb1872edfb355e12059011cdaa6ddadd04821d723cc"} Mar 13 16:10:04 crc kubenswrapper[4907]: I0313 16:10:04.595829 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:04 crc kubenswrapper[4907]: I0313 16:10:04.596163 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.136233 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.214690 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") pod \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\" (UID: \"2b3a466a-0c1c-4c86-897f-0a3375a9fb53\") " Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.223244 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw" (OuterVolumeSpecName: "kube-api-access-m4mmw") pod "2b3a466a-0c1c-4c86-897f-0a3375a9fb53" (UID: "2b3a466a-0c1c-4c86-897f-0a3375a9fb53"). InnerVolumeSpecName "kube-api-access-m4mmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.317627 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4mmw\" (UniqueName: \"kubernetes.io/projected/2b3a466a-0c1c-4c86-897f-0a3375a9fb53-kube-api-access-m4mmw\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.643266 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4vfxw" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" probeResult="failure" output=< Mar 13 16:10:05 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:10:05 crc kubenswrapper[4907]: > Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.721352 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" event={"ID":"2b3a466a-0c1c-4c86-897f-0a3375a9fb53","Type":"ContainerDied","Data":"6da23a5e86abc141883193c299ad8e19cb1959fa6ac9e21e18427d7c1e087244"} Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.721402 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6da23a5e86abc141883193c299ad8e19cb1959fa6ac9e21e18427d7c1e087244" Mar 13 16:10:05 crc kubenswrapper[4907]: I0313 16:10:05.721489 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556970-lbxp9" Mar 13 16:10:06 crc kubenswrapper[4907]: I0313 16:10:06.209932 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:10:06 crc kubenswrapper[4907]: I0313 16:10:06.221743 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556964-gfm8d"] Mar 13 16:10:07 crc kubenswrapper[4907]: I0313 16:10:07.797521 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="969a7b07-f302-48b0-b48e-730836ea94a4" path="/var/lib/kubelet/pods/969a7b07-f302-48b0-b48e-730836ea94a4/volumes" Mar 13 16:10:14 crc kubenswrapper[4907]: I0313 16:10:14.643443 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:14 crc kubenswrapper[4907]: I0313 16:10:14.703607 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:14 crc kubenswrapper[4907]: I0313 16:10:14.879804 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:10:15 crc kubenswrapper[4907]: I0313 16:10:15.823340 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4vfxw" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" containerID="cri-o://422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" gracePeriod=2 Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.308513 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.403791 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") pod \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.404186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") pod \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.404304 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") pod \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\" (UID: \"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e\") " Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.404708 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities" (OuterVolumeSpecName: "utilities") pod "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" (UID: "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.405013 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.410277 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx" (OuterVolumeSpecName: "kube-api-access-ck5cx") pod "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" (UID: "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e"). InnerVolumeSpecName "kube-api-access-ck5cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.506758 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck5cx\" (UniqueName: \"kubernetes.io/projected/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-kube-api-access-ck5cx\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.538281 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" (UID: "3b8b3456-61ff-4e08-9b0b-079a4e8fa42e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.608072 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833592 4907 generic.go:334] "Generic (PLEG): container finished" podID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" exitCode=0 Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833651 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8"} Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833693 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4vfxw" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4vfxw" event={"ID":"3b8b3456-61ff-4e08-9b0b-079a4e8fa42e","Type":"ContainerDied","Data":"c899082ad7772f673db6c6ad02b7f849f4e3f906cbceceed2c5b27f98250c37d"} Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.833739 4907 scope.go:117] "RemoveContainer" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.872925 4907 scope.go:117] "RemoveContainer" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.876740 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.892228 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4vfxw"] Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.910832 4907 scope.go:117] "RemoveContainer" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.954292 4907 scope.go:117] "RemoveContainer" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" Mar 13 16:10:16 crc kubenswrapper[4907]: E0313 16:10:16.954764 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8\": container with ID starting with 422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8 not found: ID does not exist" containerID="422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.954815 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8"} err="failed to get container status \"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8\": rpc error: code = NotFound desc = could not find container \"422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8\": container with ID starting with 422bb86eade127734f05b830061308f31ef3c58377ce0de2a7e534efc87721b8 not found: ID does not exist" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.954846 4907 scope.go:117] "RemoveContainer" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" Mar 13 16:10:16 crc kubenswrapper[4907]: E0313 16:10:16.955184 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60\": container with ID starting with 01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60 not found: ID does not exist" containerID="01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.955231 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60"} err="failed to get container status \"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60\": rpc error: code = NotFound desc = could not find container \"01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60\": container with ID starting with 01aba326819da9d4d1dbe1d2fabc474c41e125a31756f7caf14aac90810e3e60 not found: ID does not exist" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.955262 4907 scope.go:117] "RemoveContainer" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" Mar 13 16:10:16 crc kubenswrapper[4907]: E0313 16:10:16.955518 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5\": container with ID starting with 47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5 not found: ID does not exist" containerID="47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5" Mar 13 16:10:16 crc kubenswrapper[4907]: I0313 16:10:16.955548 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5"} err="failed to get container status \"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5\": rpc error: code = NotFound desc = could not find container \"47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5\": container with ID starting with 47f0591583c6efa846bfde7cb8df75e77e3b66b020180593108050ef792d97f5 not found: ID does not exist" Mar 13 16:10:17 crc kubenswrapper[4907]: I0313 16:10:17.804485 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" path="/var/lib/kubelet/pods/3b8b3456-61ff-4e08-9b0b-079a4e8fa42e/volumes" Mar 13 16:10:18 crc kubenswrapper[4907]: I0313 16:10:18.860847 4907 generic.go:334] "Generic (PLEG): container finished" podID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerID="af77d9f29bb5e409e89eed0618c3878a4fc70669c940d590206ec36dfdf086f6" exitCode=0 Mar 13 16:10:18 crc kubenswrapper[4907]: I0313 16:10:18.860967 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerDied","Data":"af77d9f29bb5e409e89eed0618c3878a4fc70669c940d590206ec36dfdf086f6"} Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.351908 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394207 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394308 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394386 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.394446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") pod \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\" (UID: \"bfe0612b-2d05-4625-b14e-8b9ea02a8c06\") " Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.400752 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p" (OuterVolumeSpecName: "kube-api-access-gcn6p") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "kube-api-access-gcn6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.400972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph" (OuterVolumeSpecName: "ceph") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.427164 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory" (OuterVolumeSpecName: "inventory") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.438629 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "bfe0612b-2d05-4625-b14e-8b9ea02a8c06" (UID: "bfe0612b-2d05-4625-b14e-8b9ea02a8c06"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497074 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497206 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497290 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.497350 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcn6p\" (UniqueName: \"kubernetes.io/projected/bfe0612b-2d05-4625-b14e-8b9ea02a8c06-kube-api-access-gcn6p\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.882160 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" event={"ID":"bfe0612b-2d05-4625-b14e-8b9ea02a8c06","Type":"ContainerDied","Data":"ff5e634a1d2fde12b87f2739aa06c6e49877afb509500e105fbf347bc7380990"} Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.882234 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff5e634a1d2fde12b87f2739aa06c6e49877afb509500e105fbf347bc7380990" Mar 13 16:10:20 crc kubenswrapper[4907]: I0313 16:10:20.882323 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-wrvtc" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.016639 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-rrpt2"] Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017460 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-utilities" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017480 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-utilities" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017516 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017524 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017558 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-content" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017566 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="extract-content" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017584 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerName="oc" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017593 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerName="oc" Mar 13 16:10:21 crc kubenswrapper[4907]: E0313 16:10:21.017606 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerName="configure-network-openstack-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017614 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerName="configure-network-openstack-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017867 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8b3456-61ff-4e08-9b0b-079a4e8fa42e" containerName="registry-server" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017905 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" containerName="oc" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.017919 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe0612b-2d05-4625-b14e-8b9ea02a8c06" containerName="configure-network-openstack-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.018755 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.021867 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.022584 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.022849 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.023110 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.027911 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.027963 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.028008 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.028075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.029303 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-rrpt2"] Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.129482 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.129574 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.130137 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.130257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.135037 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.135579 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.136073 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.146847 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"validate-network-openstack-openstack-cell1-rrpt2\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.337135 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:21 crc kubenswrapper[4907]: I0313 16:10:21.931736 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-rrpt2"] Mar 13 16:10:21 crc kubenswrapper[4907]: W0313 16:10:21.941038 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a4f6803_5919_4737_acd0_5b25e12a7c55.slice/crio-f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6 WatchSource:0}: Error finding container f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6: Status 404 returned error can't find the container with id f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6 Mar 13 16:10:22 crc kubenswrapper[4907]: I0313 16:10:22.909562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerStarted","Data":"f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6"} Mar 13 16:10:23 crc kubenswrapper[4907]: I0313 16:10:23.919239 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerStarted","Data":"3982867b847d25c43ce67d6447d7efe49307a957e163fd43946487b24dfbdbe8"} Mar 13 16:10:23 crc kubenswrapper[4907]: I0313 16:10:23.942728 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" podStartSLOduration=3.241210455 podStartE2EDuration="3.942708365s" podCreationTimestamp="2026-03-13 16:10:20 +0000 UTC" firstStartedPulling="2026-03-13 16:10:21.942910181 +0000 UTC m=+7520.842697860" lastFinishedPulling="2026-03-13 16:10:22.644408071 +0000 UTC m=+7521.544195770" observedRunningTime="2026-03-13 16:10:23.933970046 +0000 UTC m=+7522.833757735" watchObservedRunningTime="2026-03-13 16:10:23.942708365 +0000 UTC m=+7522.842496054" Mar 13 16:10:28 crc kubenswrapper[4907]: I0313 16:10:28.967153 4907 generic.go:334] "Generic (PLEG): container finished" podID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerID="3982867b847d25c43ce67d6447d7efe49307a957e163fd43946487b24dfbdbe8" exitCode=0 Mar 13 16:10:28 crc kubenswrapper[4907]: I0313 16:10:28.967264 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerDied","Data":"3982867b847d25c43ce67d6447d7efe49307a957e163fd43946487b24dfbdbe8"} Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.442475 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551436 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551510 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551624 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.551666 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") pod \"9a4f6803-5919-4737-acd0-5b25e12a7c55\" (UID: \"9a4f6803-5919-4737-acd0-5b25e12a7c55\") " Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.556390 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t" (OuterVolumeSpecName: "kube-api-access-nb44t") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "kube-api-access-nb44t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.558482 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph" (OuterVolumeSpecName: "ceph") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.586286 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory" (OuterVolumeSpecName: "inventory") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.591020 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "9a4f6803-5919-4737-acd0-5b25e12a7c55" (UID: "9a4f6803-5919-4737-acd0-5b25e12a7c55"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654272 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654306 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654316 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9a4f6803-5919-4737-acd0-5b25e12a7c55-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:30 crc kubenswrapper[4907]: I0313 16:10:30.654325 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb44t\" (UniqueName: \"kubernetes.io/projected/9a4f6803-5919-4737-acd0-5b25e12a7c55-kube-api-access-nb44t\") on node \"crc\" DevicePath \"\"" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.024391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" event={"ID":"9a4f6803-5919-4737-acd0-5b25e12a7c55","Type":"ContainerDied","Data":"f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6"} Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.024435 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6512817fa65e9fa0d8596c44a85970e3319de57228fe970260d0992907856d6" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.024465 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-rrpt2" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.099575 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vw9h5"] Mar 13 16:10:31 crc kubenswrapper[4907]: E0313 16:10:31.100090 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerName="validate-network-openstack-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.100111 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerName="validate-network-openstack-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.100311 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4f6803-5919-4737-acd0-5b25e12a7c55" containerName="validate-network-openstack-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.101072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.105799 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.105794 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.106282 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.106401 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.156996 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vw9h5"] Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.267999 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.268392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.268497 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.268634 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.370899 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.371448 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.371515 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.371563 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.383932 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.384292 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.384825 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.396786 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"install-os-openstack-openstack-cell1-vw9h5\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.424836 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:10:31 crc kubenswrapper[4907]: I0313 16:10:31.999097 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-vw9h5"] Mar 13 16:10:32 crc kubenswrapper[4907]: I0313 16:10:32.052137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerStarted","Data":"4eeb9f6355cdd98a6c410784295be2591cbbb0e5ae4d88b6b06b43662075461d"} Mar 13 16:10:33 crc kubenswrapper[4907]: I0313 16:10:33.062776 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerStarted","Data":"86fb48771b3fd8ed723bd1ee960c478c1f6aca7779807b8ba2adcc20af416927"} Mar 13 16:10:33 crc kubenswrapper[4907]: I0313 16:10:33.086950 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" podStartSLOduration=1.601636751 podStartE2EDuration="2.086932172s" podCreationTimestamp="2026-03-13 16:10:31 +0000 UTC" firstStartedPulling="2026-03-13 16:10:32.005521816 +0000 UTC m=+7530.905309495" lastFinishedPulling="2026-03-13 16:10:32.490817237 +0000 UTC m=+7531.390604916" observedRunningTime="2026-03-13 16:10:33.079742435 +0000 UTC m=+7531.979530134" watchObservedRunningTime="2026-03-13 16:10:33.086932172 +0000 UTC m=+7531.986719861" Mar 13 16:10:41 crc kubenswrapper[4907]: I0313 16:10:41.281575 4907 scope.go:117] "RemoveContainer" containerID="e87bfe7ae00d8363f28ac3bbec93999ba43f3034b744a941e08c8397ba4ba459" Mar 13 16:11:19 crc kubenswrapper[4907]: I0313 16:11:19.565108 4907 generic.go:334] "Generic (PLEG): container finished" podID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerID="86fb48771b3fd8ed723bd1ee960c478c1f6aca7779807b8ba2adcc20af416927" exitCode=0 Mar 13 16:11:19 crc kubenswrapper[4907]: I0313 16:11:19.565181 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerDied","Data":"86fb48771b3fd8ed723bd1ee960c478c1f6aca7779807b8ba2adcc20af416927"} Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.115157 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219797 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219958 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.219994 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") pod \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\" (UID: \"cc30c8d7-5f7e-47c8-b89f-6a638b886769\") " Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.228073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf" (OuterVolumeSpecName: "kube-api-access-p22zf") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "kube-api-access-p22zf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.234566 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph" (OuterVolumeSpecName: "ceph") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.251334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.275968 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory" (OuterVolumeSpecName: "inventory") pod "cc30c8d7-5f7e-47c8-b89f-6a638b886769" (UID: "cc30c8d7-5f7e-47c8-b89f-6a638b886769"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323558 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p22zf\" (UniqueName: \"kubernetes.io/projected/cc30c8d7-5f7e-47c8-b89f-6a638b886769-kube-api-access-p22zf\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323621 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323640 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.323659 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cc30c8d7-5f7e-47c8-b89f-6a638b886769-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.588390 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" event={"ID":"cc30c8d7-5f7e-47c8-b89f-6a638b886769","Type":"ContainerDied","Data":"4eeb9f6355cdd98a6c410784295be2591cbbb0e5ae4d88b6b06b43662075461d"} Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.588748 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eeb9f6355cdd98a6c410784295be2591cbbb0e5ae4d88b6b06b43662075461d" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.588542 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-vw9h5" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.690756 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s9llf"] Mar 13 16:11:21 crc kubenswrapper[4907]: E0313 16:11:21.691703 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerName="install-os-openstack-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.691825 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerName="install-os-openstack-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.692265 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc30c8d7-5f7e-47c8-b89f-6a638b886769" containerName="install-os-openstack-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.693617 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.696755 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.697753 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.698151 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.702847 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s9llf"] Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.707503 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.834771 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.835052 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.835094 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.835370 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938077 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938209 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938227 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.938365 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.943962 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.944030 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.945045 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:21 crc kubenswrapper[4907]: I0313 16:11:21.960561 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"configure-os-openstack-openstack-cell1-s9llf\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.011176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.600017 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-s9llf"] Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.609134 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.717604 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.720810 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.741796 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.868840 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.869604 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.869646 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972474 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972573 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.972909 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.973520 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:22 crc kubenswrapper[4907]: I0313 16:11:22.999636 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"certified-operators-rq8fl\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.045475 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.553400 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:23 crc kubenswrapper[4907]: W0313 16:11:23.557767 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda822760f_41c5_40ab_9b8a_64f4a465c25f.slice/crio-bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde WatchSource:0}: Error finding container bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde: Status 404 returned error can't find the container with id bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.616806 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerStarted","Data":"5ea72312e8819f02cc06ddf2ae911b43e3f99fe396d78ccee0fc42da52e0192d"} Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.616854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerStarted","Data":"463d170ab3d7d3c1c1ed4d54bd7ad5aae875c69e0edf649f17b5303e58e8fd44"} Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.619323 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerStarted","Data":"bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde"} Mar 13 16:11:23 crc kubenswrapper[4907]: I0313 16:11:23.647347 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" podStartSLOduration=1.87948698 podStartE2EDuration="2.647327813s" podCreationTimestamp="2026-03-13 16:11:21 +0000 UTC" firstStartedPulling="2026-03-13 16:11:22.608849102 +0000 UTC m=+7581.508636791" lastFinishedPulling="2026-03-13 16:11:23.376689945 +0000 UTC m=+7582.276477624" observedRunningTime="2026-03-13 16:11:23.640607359 +0000 UTC m=+7582.540395048" watchObservedRunningTime="2026-03-13 16:11:23.647327813 +0000 UTC m=+7582.547115502" Mar 13 16:11:24 crc kubenswrapper[4907]: I0313 16:11:24.629983 4907 generic.go:334] "Generic (PLEG): container finished" podID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" exitCode=0 Mar 13 16:11:24 crc kubenswrapper[4907]: I0313 16:11:24.630043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9"} Mar 13 16:11:26 crc kubenswrapper[4907]: I0313 16:11:26.653083 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerStarted","Data":"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58"} Mar 13 16:11:28 crc kubenswrapper[4907]: I0313 16:11:28.670124 4907 generic.go:334] "Generic (PLEG): container finished" podID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" exitCode=0 Mar 13 16:11:28 crc kubenswrapper[4907]: I0313 16:11:28.670186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58"} Mar 13 16:11:29 crc kubenswrapper[4907]: I0313 16:11:29.681614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerStarted","Data":"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3"} Mar 13 16:11:29 crc kubenswrapper[4907]: I0313 16:11:29.711501 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rq8fl" podStartSLOduration=3.225665568 podStartE2EDuration="7.711482796s" podCreationTimestamp="2026-03-13 16:11:22 +0000 UTC" firstStartedPulling="2026-03-13 16:11:24.632508704 +0000 UTC m=+7583.532296393" lastFinishedPulling="2026-03-13 16:11:29.118325932 +0000 UTC m=+7588.018113621" observedRunningTime="2026-03-13 16:11:29.707292281 +0000 UTC m=+7588.607079970" watchObservedRunningTime="2026-03-13 16:11:29.711482796 +0000 UTC m=+7588.611270485" Mar 13 16:11:33 crc kubenswrapper[4907]: I0313 16:11:33.046242 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:33 crc kubenswrapper[4907]: I0313 16:11:33.046768 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:33 crc kubenswrapper[4907]: I0313 16:11:33.124026 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:43 crc kubenswrapper[4907]: I0313 16:11:43.155640 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:43 crc kubenswrapper[4907]: I0313 16:11:43.224188 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:43 crc kubenswrapper[4907]: I0313 16:11:43.811728 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rq8fl" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" containerID="cri-o://c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" gracePeriod=2 Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.755583 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827657 4907 generic.go:334] "Generic (PLEG): container finished" podID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" exitCode=0 Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827723 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3"} Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827820 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rq8fl" event={"ID":"a822760f-41c5-40ab-9b8a-64f4a465c25f","Type":"ContainerDied","Data":"bdc94ca7df4e674f0317453fa5febaf6716d3284c442d035f275bc10a9e3bbde"} Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827835 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rq8fl" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.827854 4907 scope.go:117] "RemoveContainer" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.861979 4907 scope.go:117] "RemoveContainer" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.870361 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") pod \"a822760f-41c5-40ab-9b8a-64f4a465c25f\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.870676 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") pod \"a822760f-41c5-40ab-9b8a-64f4a465c25f\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.870728 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") pod \"a822760f-41c5-40ab-9b8a-64f4a465c25f\" (UID: \"a822760f-41c5-40ab-9b8a-64f4a465c25f\") " Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.871725 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities" (OuterVolumeSpecName: "utilities") pod "a822760f-41c5-40ab-9b8a-64f4a465c25f" (UID: "a822760f-41c5-40ab-9b8a-64f4a465c25f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.881734 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6" (OuterVolumeSpecName: "kube-api-access-g2jm6") pod "a822760f-41c5-40ab-9b8a-64f4a465c25f" (UID: "a822760f-41c5-40ab-9b8a-64f4a465c25f"). InnerVolumeSpecName "kube-api-access-g2jm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.890014 4907 scope.go:117] "RemoveContainer" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.972764 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a822760f-41c5-40ab-9b8a-64f4a465c25f" (UID: "a822760f-41c5-40ab-9b8a-64f4a465c25f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.973475 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2jm6\" (UniqueName: \"kubernetes.io/projected/a822760f-41c5-40ab-9b8a-64f4a465c25f-kube-api-access-g2jm6\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.973510 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:44 crc kubenswrapper[4907]: I0313 16:11:44.973520 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a822760f-41c5-40ab-9b8a-64f4a465c25f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.009655 4907 scope.go:117] "RemoveContainer" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" Mar 13 16:11:45 crc kubenswrapper[4907]: E0313 16:11:45.010316 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3\": container with ID starting with c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3 not found: ID does not exist" containerID="c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010361 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3"} err="failed to get container status \"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3\": rpc error: code = NotFound desc = could not find container \"c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3\": container with ID starting with c2f9dccdffbee706cf1dccbed6bcf5f01e43c63bb6ec1faaef0a1728b7246db3 not found: ID does not exist" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010395 4907 scope.go:117] "RemoveContainer" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" Mar 13 16:11:45 crc kubenswrapper[4907]: E0313 16:11:45.010797 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58\": container with ID starting with 1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58 not found: ID does not exist" containerID="1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010826 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58"} err="failed to get container status \"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58\": rpc error: code = NotFound desc = could not find container \"1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58\": container with ID starting with 1d61ea08412ba1da32fdc2116795bc5b58af53fd12a07f6d91a43af634811d58 not found: ID does not exist" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.010863 4907 scope.go:117] "RemoveContainer" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" Mar 13 16:11:45 crc kubenswrapper[4907]: E0313 16:11:45.011268 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9\": container with ID starting with eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9 not found: ID does not exist" containerID="eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.011303 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9"} err="failed to get container status \"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9\": rpc error: code = NotFound desc = could not find container \"eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9\": container with ID starting with eb08050c707ce49c267ab4b19c8d14852dd33b822d7c83e2aa9e2003f02b3fd9 not found: ID does not exist" Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.165923 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.175377 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rq8fl"] Mar 13 16:11:45 crc kubenswrapper[4907]: I0313 16:11:45.824364 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" path="/var/lib/kubelet/pods/a822760f-41c5-40ab-9b8a-64f4a465c25f/volumes" Mar 13 16:11:48 crc kubenswrapper[4907]: I0313 16:11:48.041265 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:11:48 crc kubenswrapper[4907]: I0313 16:11:48.041839 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.177355 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:12:00 crc kubenswrapper[4907]: E0313 16:12:00.178591 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-utilities" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.178605 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-utilities" Mar 13 16:12:00 crc kubenswrapper[4907]: E0313 16:12:00.178640 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-content" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.178648 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="extract-content" Mar 13 16:12:00 crc kubenswrapper[4907]: E0313 16:12:00.178677 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.178684 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.179245 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="a822760f-41c5-40ab-9b8a-64f4a465c25f" containerName="registry-server" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.184740 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.186922 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.187413 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.188457 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.192319 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.353854 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"auto-csr-approver-29556972-llpxn\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.455622 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"auto-csr-approver-29556972-llpxn\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.475293 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"auto-csr-approver-29556972-llpxn\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.506647 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:00 crc kubenswrapper[4907]: I0313 16:12:00.977774 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:12:01 crc kubenswrapper[4907]: I0313 16:12:01.988635 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556972-llpxn" event={"ID":"85500c41-93c1-44ae-b418-d7f125c04621","Type":"ContainerStarted","Data":"aa6902af445c9119dffb6494cc2d7852bdb913b5b91041562d50c8af3f525493"} Mar 13 16:12:02 crc kubenswrapper[4907]: I0313 16:12:02.998853 4907 generic.go:334] "Generic (PLEG): container finished" podID="85500c41-93c1-44ae-b418-d7f125c04621" containerID="1597ac303cc302238cff89d1286c78c824124bc23ca555a00a5dd03f41ac4104" exitCode=0 Mar 13 16:12:02 crc kubenswrapper[4907]: I0313 16:12:02.998933 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556972-llpxn" event={"ID":"85500c41-93c1-44ae-b418-d7f125c04621","Type":"ContainerDied","Data":"1597ac303cc302238cff89d1286c78c824124bc23ca555a00a5dd03f41ac4104"} Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.471116 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.560180 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") pod \"85500c41-93c1-44ae-b418-d7f125c04621\" (UID: \"85500c41-93c1-44ae-b418-d7f125c04621\") " Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.566677 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8" (OuterVolumeSpecName: "kube-api-access-rmjn8") pod "85500c41-93c1-44ae-b418-d7f125c04621" (UID: "85500c41-93c1-44ae-b418-d7f125c04621"). InnerVolumeSpecName "kube-api-access-rmjn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:04 crc kubenswrapper[4907]: I0313 16:12:04.663610 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmjn8\" (UniqueName: \"kubernetes.io/projected/85500c41-93c1-44ae-b418-d7f125c04621-kube-api-access-rmjn8\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.023935 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556972-llpxn" event={"ID":"85500c41-93c1-44ae-b418-d7f125c04621","Type":"ContainerDied","Data":"aa6902af445c9119dffb6494cc2d7852bdb913b5b91041562d50c8af3f525493"} Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.023973 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa6902af445c9119dffb6494cc2d7852bdb913b5b91041562d50c8af3f525493" Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.023979 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556972-llpxn" Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.546718 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.560828 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556966-drz8k"] Mar 13 16:12:05 crc kubenswrapper[4907]: I0313 16:12:05.796848 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6bd258-54f3-49e8-8b92-d101e6b19872" path="/var/lib/kubelet/pods/9e6bd258-54f3-49e8-8b92-d101e6b19872/volumes" Mar 13 16:12:09 crc kubenswrapper[4907]: I0313 16:12:09.065111 4907 generic.go:334] "Generic (PLEG): container finished" podID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerID="5ea72312e8819f02cc06ddf2ae911b43e3f99fe396d78ccee0fc42da52e0192d" exitCode=0 Mar 13 16:12:09 crc kubenswrapper[4907]: I0313 16:12:09.065230 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerDied","Data":"5ea72312e8819f02cc06ddf2ae911b43e3f99fe396d78ccee0fc42da52e0192d"} Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.625555 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818472 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818536 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818691 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.818727 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") pod \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\" (UID: \"173f7de2-ef0c-4750-a8b3-23e078ab9bde\") " Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.833031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph" (OuterVolumeSpecName: "ceph") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.834412 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f" (OuterVolumeSpecName: "kube-api-access-mt98f") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "kube-api-access-mt98f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.846343 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory" (OuterVolumeSpecName: "inventory") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.862466 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "173f7de2-ef0c-4750-a8b3-23e078ab9bde" (UID: "173f7de2-ef0c-4750-a8b3-23e078ab9bde"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921419 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921451 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt98f\" (UniqueName: \"kubernetes.io/projected/173f7de2-ef0c-4750-a8b3-23e078ab9bde-kube-api-access-mt98f\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:10 crc kubenswrapper[4907]: I0313 16:12:10.921473 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/173f7de2-ef0c-4750-a8b3-23e078ab9bde-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.085383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" event={"ID":"173f7de2-ef0c-4750-a8b3-23e078ab9bde","Type":"ContainerDied","Data":"463d170ab3d7d3c1c1ed4d54bd7ad5aae875c69e0edf649f17b5303e58e8fd44"} Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.085432 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="463d170ab3d7d3c1c1ed4d54bd7ad5aae875c69e0edf649f17b5303e58e8fd44" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.085432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-s9llf" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.178297 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-phzr9"] Mar 13 16:12:11 crc kubenswrapper[4907]: E0313 16:12:11.179045 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85500c41-93c1-44ae-b418-d7f125c04621" containerName="oc" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179074 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="85500c41-93c1-44ae-b418-d7f125c04621" containerName="oc" Mar 13 16:12:11 crc kubenswrapper[4907]: E0313 16:12:11.179093 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerName="configure-os-openstack-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179101 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerName="configure-os-openstack-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179309 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="85500c41-93c1-44ae-b418-d7f125c04621" containerName="oc" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.179330 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="173f7de2-ef0c-4750-a8b3-23e078ab9bde" containerName="configure-os-openstack-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.180181 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.182586 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.182785 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.183239 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.184279 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.219157 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-phzr9"] Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331189 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331380 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331582 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.331758 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.434582 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.434723 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.434918 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.435035 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.444174 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.444415 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.458805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.468044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"ssh-known-hosts-openstack-phzr9\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:11 crc kubenswrapper[4907]: I0313 16:12:11.501640 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:12 crc kubenswrapper[4907]: I0313 16:12:12.140164 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-phzr9"] Mar 13 16:12:13 crc kubenswrapper[4907]: I0313 16:12:13.127681 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerStarted","Data":"fae553dd1e8789a4971d11997504d70757ec24d649cdf5a39afeb67e1286709d"} Mar 13 16:12:13 crc kubenswrapper[4907]: I0313 16:12:13.128267 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerStarted","Data":"3606c840833edcd74a624c6afbba128aa52c01cc3569fec6423293e6c8311b53"} Mar 13 16:12:13 crc kubenswrapper[4907]: I0313 16:12:13.158499 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-phzr9" podStartSLOduration=1.679763903 podStartE2EDuration="2.158478044s" podCreationTimestamp="2026-03-13 16:12:11 +0000 UTC" firstStartedPulling="2026-03-13 16:12:12.161353065 +0000 UTC m=+7631.061140774" lastFinishedPulling="2026-03-13 16:12:12.640067216 +0000 UTC m=+7631.539854915" observedRunningTime="2026-03-13 16:12:13.144360077 +0000 UTC m=+7632.044147766" watchObservedRunningTime="2026-03-13 16:12:13.158478044 +0000 UTC m=+7632.058265733" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.672519 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.675236 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.716038 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.730954 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.731017 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.731374 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.832355 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.832416 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.832538 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.833000 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.833136 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:15 crc kubenswrapper[4907]: I0313 16:12:15.851519 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"community-operators-9q59x\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:16 crc kubenswrapper[4907]: I0313 16:12:16.018645 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:16 crc kubenswrapper[4907]: I0313 16:12:16.571496 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:17 crc kubenswrapper[4907]: I0313 16:12:17.165739 4907 generic.go:334] "Generic (PLEG): container finished" podID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" exitCode=0 Mar 13 16:12:17 crc kubenswrapper[4907]: I0313 16:12:17.165784 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535"} Mar 13 16:12:17 crc kubenswrapper[4907]: I0313 16:12:17.166016 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerStarted","Data":"e562e09fb98e987deb99f7ca2ebea739da366ebb50fdb52113a88f14af75db5b"} Mar 13 16:12:18 crc kubenswrapper[4907]: I0313 16:12:18.042014 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:12:18 crc kubenswrapper[4907]: I0313 16:12:18.042711 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:12:18 crc kubenswrapper[4907]: I0313 16:12:18.194557 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerStarted","Data":"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02"} Mar 13 16:12:20 crc kubenswrapper[4907]: I0313 16:12:20.218763 4907 generic.go:334] "Generic (PLEG): container finished" podID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" exitCode=0 Mar 13 16:12:20 crc kubenswrapper[4907]: I0313 16:12:20.218836 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02"} Mar 13 16:12:21 crc kubenswrapper[4907]: I0313 16:12:21.231249 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerStarted","Data":"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445"} Mar 13 16:12:21 crc kubenswrapper[4907]: I0313 16:12:21.266370 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9q59x" podStartSLOduration=2.477429075 podStartE2EDuration="6.266354991s" podCreationTimestamp="2026-03-13 16:12:15 +0000 UTC" firstStartedPulling="2026-03-13 16:12:17.169039535 +0000 UTC m=+7636.068827224" lastFinishedPulling="2026-03-13 16:12:20.957965441 +0000 UTC m=+7639.857753140" observedRunningTime="2026-03-13 16:12:21.2597576 +0000 UTC m=+7640.159545289" watchObservedRunningTime="2026-03-13 16:12:21.266354991 +0000 UTC m=+7640.166142680" Mar 13 16:12:22 crc kubenswrapper[4907]: I0313 16:12:22.239762 4907 generic.go:334] "Generic (PLEG): container finished" podID="4aef751c-592b-403e-b581-92c88cf65785" containerID="fae553dd1e8789a4971d11997504d70757ec24d649cdf5a39afeb67e1286709d" exitCode=0 Mar 13 16:12:22 crc kubenswrapper[4907]: I0313 16:12:22.239875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerDied","Data":"fae553dd1e8789a4971d11997504d70757ec24d649cdf5a39afeb67e1286709d"} Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.845314 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921525 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921617 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921675 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.921711 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") pod \"4aef751c-592b-403e-b581-92c88cf65785\" (UID: \"4aef751c-592b-403e-b581-92c88cf65785\") " Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.930583 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq" (OuterVolumeSpecName: "kube-api-access-tg6tq") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "kube-api-access-tg6tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.931188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph" (OuterVolumeSpecName: "ceph") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.949856 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:23 crc kubenswrapper[4907]: I0313 16:12:23.950268 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "4aef751c-592b-403e-b581-92c88cf65785" (UID: "4aef751c-592b-403e-b581-92c88cf65785"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024066 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg6tq\" (UniqueName: \"kubernetes.io/projected/4aef751c-592b-403e-b581-92c88cf65785-kube-api-access-tg6tq\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024096 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024106 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.024116 4907 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/4aef751c-592b-403e-b581-92c88cf65785-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.260800 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-phzr9" event={"ID":"4aef751c-592b-403e-b581-92c88cf65785","Type":"ContainerDied","Data":"3606c840833edcd74a624c6afbba128aa52c01cc3569fec6423293e6c8311b53"} Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.260840 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3606c840833edcd74a624c6afbba128aa52c01cc3569fec6423293e6c8311b53" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.260916 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-phzr9" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.364673 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-hml5k"] Mar 13 16:12:24 crc kubenswrapper[4907]: E0313 16:12:24.365209 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aef751c-592b-403e-b581-92c88cf65785" containerName="ssh-known-hosts-openstack" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.365227 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aef751c-592b-403e-b581-92c88cf65785" containerName="ssh-known-hosts-openstack" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.365443 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aef751c-592b-403e-b581-92c88cf65785" containerName="ssh-known-hosts-openstack" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.366239 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.370252 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.372735 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.372810 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.372820 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.378947 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-hml5k"] Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434097 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434302 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434339 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.434549 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536520 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536559 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536626 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.536686 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.541541 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.546569 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.554016 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.558627 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"run-os-openstack-openstack-cell1-hml5k\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:24 crc kubenswrapper[4907]: I0313 16:12:24.700298 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:25 crc kubenswrapper[4907]: I0313 16:12:25.249993 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-hml5k"] Mar 13 16:12:25 crc kubenswrapper[4907]: W0313 16:12:25.250060 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34db6153_7522_4aae_9712_eb87cc9aa543.slice/crio-1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e WatchSource:0}: Error finding container 1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e: Status 404 returned error can't find the container with id 1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e Mar 13 16:12:25 crc kubenswrapper[4907]: I0313 16:12:25.278245 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerStarted","Data":"1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e"} Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.019855 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.020178 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.068530 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.291050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerStarted","Data":"1a4edf4adc052232eadad9bc1d0620ff53938e2a203a59f9c95cd4e2f267081a"} Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.336007 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.365167 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-hml5k" podStartSLOduration=1.864461102 podStartE2EDuration="2.365144924s" podCreationTimestamp="2026-03-13 16:12:24 +0000 UTC" firstStartedPulling="2026-03-13 16:12:25.252453403 +0000 UTC m=+7644.152241102" lastFinishedPulling="2026-03-13 16:12:25.753137195 +0000 UTC m=+7644.652924924" observedRunningTime="2026-03-13 16:12:26.313428789 +0000 UTC m=+7645.213216488" watchObservedRunningTime="2026-03-13 16:12:26.365144924 +0000 UTC m=+7645.264932623" Mar 13 16:12:26 crc kubenswrapper[4907]: I0313 16:12:26.386311 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:28 crc kubenswrapper[4907]: I0313 16:12:28.307693 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9q59x" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" containerID="cri-o://d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" gracePeriod=2 Mar 13 16:12:28 crc kubenswrapper[4907]: I0313 16:12:28.899351 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.039521 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") pod \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.039598 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") pod \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.039649 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") pod \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\" (UID: \"aabfe118-3ad6-4625-bdab-6a1f2bc50890\") " Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.040899 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities" (OuterVolumeSpecName: "utilities") pod "aabfe118-3ad6-4625-bdab-6a1f2bc50890" (UID: "aabfe118-3ad6-4625-bdab-6a1f2bc50890"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.054229 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b" (OuterVolumeSpecName: "kube-api-access-9xr6b") pod "aabfe118-3ad6-4625-bdab-6a1f2bc50890" (UID: "aabfe118-3ad6-4625-bdab-6a1f2bc50890"). InnerVolumeSpecName "kube-api-access-9xr6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.092629 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aabfe118-3ad6-4625-bdab-6a1f2bc50890" (UID: "aabfe118-3ad6-4625-bdab-6a1f2bc50890"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.142053 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.142094 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xr6b\" (UniqueName: \"kubernetes.io/projected/aabfe118-3ad6-4625-bdab-6a1f2bc50890-kube-api-access-9xr6b\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.142108 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aabfe118-3ad6-4625-bdab-6a1f2bc50890-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318326 4907 generic.go:334] "Generic (PLEG): container finished" podID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" exitCode=0 Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318384 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445"} Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318415 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9q59x" event={"ID":"aabfe118-3ad6-4625-bdab-6a1f2bc50890","Type":"ContainerDied","Data":"e562e09fb98e987deb99f7ca2ebea739da366ebb50fdb52113a88f14af75db5b"} Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318434 4907 scope.go:117] "RemoveContainer" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.318426 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9q59x" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.350832 4907 scope.go:117] "RemoveContainer" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.351487 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.363065 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9q59x"] Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.375650 4907 scope.go:117] "RemoveContainer" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.437717 4907 scope.go:117] "RemoveContainer" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" Mar 13 16:12:29 crc kubenswrapper[4907]: E0313 16:12:29.438984 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445\": container with ID starting with d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445 not found: ID does not exist" containerID="d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.439045 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445"} err="failed to get container status \"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445\": rpc error: code = NotFound desc = could not find container \"d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445\": container with ID starting with d01c41ac19b3db32d5d13701af11e1e2508fcdc64db4c92f02188f251ddd2445 not found: ID does not exist" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.439088 4907 scope.go:117] "RemoveContainer" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" Mar 13 16:12:29 crc kubenswrapper[4907]: E0313 16:12:29.439947 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02\": container with ID starting with c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02 not found: ID does not exist" containerID="c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.440000 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02"} err="failed to get container status \"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02\": rpc error: code = NotFound desc = could not find container \"c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02\": container with ID starting with c706458164f7c6287d29e34e31517532fdb1cc00947568aaa9db8fa03c2c9c02 not found: ID does not exist" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.440018 4907 scope.go:117] "RemoveContainer" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" Mar 13 16:12:29 crc kubenswrapper[4907]: E0313 16:12:29.440613 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535\": container with ID starting with 4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535 not found: ID does not exist" containerID="4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.440644 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535"} err="failed to get container status \"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535\": rpc error: code = NotFound desc = could not find container \"4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535\": container with ID starting with 4ec6ccaa677ccbbb66f21ece73e47ce24307c31346e8c97bf83dbdeb0ab4c535 not found: ID does not exist" Mar 13 16:12:29 crc kubenswrapper[4907]: I0313 16:12:29.793459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" path="/var/lib/kubelet/pods/aabfe118-3ad6-4625-bdab-6a1f2bc50890/volumes" Mar 13 16:12:36 crc kubenswrapper[4907]: I0313 16:12:36.390980 4907 generic.go:334] "Generic (PLEG): container finished" podID="34db6153-7522-4aae-9712-eb87cc9aa543" containerID="1a4edf4adc052232eadad9bc1d0620ff53938e2a203a59f9c95cd4e2f267081a" exitCode=0 Mar 13 16:12:36 crc kubenswrapper[4907]: I0313 16:12:36.391050 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerDied","Data":"1a4edf4adc052232eadad9bc1d0620ff53938e2a203a59f9c95cd4e2f267081a"} Mar 13 16:12:37 crc kubenswrapper[4907]: I0313 16:12:37.926535 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.042902 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.043070 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.043155 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.043186 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") pod \"34db6153-7522-4aae-9712-eb87cc9aa543\" (UID: \"34db6153-7522-4aae-9712-eb87cc9aa543\") " Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.052166 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf" (OuterVolumeSpecName: "kube-api-access-7dstf") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "kube-api-access-7dstf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.065168 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph" (OuterVolumeSpecName: "ceph") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.080294 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory" (OuterVolumeSpecName: "inventory") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.087212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "34db6153-7522-4aae-9712-eb87cc9aa543" (UID: "34db6153-7522-4aae-9712-eb87cc9aa543"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146153 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dstf\" (UniqueName: \"kubernetes.io/projected/34db6153-7522-4aae-9712-eb87cc9aa543-kube-api-access-7dstf\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146554 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146575 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.146596 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/34db6153-7522-4aae-9712-eb87cc9aa543-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.419098 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-hml5k" event={"ID":"34db6153-7522-4aae-9712-eb87cc9aa543","Type":"ContainerDied","Data":"1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e"} Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.419161 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1120479ce59841724e323313ca8d2a10c18000da39d4ac71ea46d45f1179803e" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.419195 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-hml5k" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.482642 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x7hvj"] Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483110 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-content" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483124 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-content" Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483146 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-utilities" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483151 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="extract-utilities" Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483176 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34db6153-7522-4aae-9712-eb87cc9aa543" containerName="run-os-openstack-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483181 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="34db6153-7522-4aae-9712-eb87cc9aa543" containerName="run-os-openstack-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: E0313 16:12:38.483193 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483199 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483388 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="aabfe118-3ad6-4625-bdab-6a1f2bc50890" containerName="registry-server" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.483406 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="34db6153-7522-4aae-9712-eb87cc9aa543" containerName="run-os-openstack-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.484177 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.489479 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.489723 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.489851 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.490246 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.495106 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x7hvj"] Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554114 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554208 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.554721 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.657668 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.658192 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.658322 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.658419 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.664210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.664611 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.666687 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.681653 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"reboot-os-openstack-openstack-cell1-x7hvj\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:38 crc kubenswrapper[4907]: I0313 16:12:38.803176 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:39 crc kubenswrapper[4907]: I0313 16:12:39.326493 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x7hvj"] Mar 13 16:12:39 crc kubenswrapper[4907]: I0313 16:12:39.428753 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerStarted","Data":"895e8f2c18c5de20a20d4cb067c8d1021996b85a83d2443bfe64b2080b187ce0"} Mar 13 16:12:40 crc kubenswrapper[4907]: I0313 16:12:40.443257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerStarted","Data":"84b83daed0821729d3e57629749b6c8b5286ebc80f25b88784c7bd96792460bd"} Mar 13 16:12:41 crc kubenswrapper[4907]: I0313 16:12:41.427066 4907 scope.go:117] "RemoveContainer" containerID="fba4ffaa28135dbeabe5c056323b0e99fffec19d264cb5778eaf0cb6c3cc2d40" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.041516 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.041961 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.042007 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.042789 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.042843 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560" gracePeriod=600 Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522093 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560" exitCode=0 Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522580 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560"} Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522679 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f"} Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.522708 4907 scope.go:117] "RemoveContainer" containerID="97ad1b399368adb5c1fc3c4cf2810fa240ec320625c36e99f246d6171b951964" Mar 13 16:12:48 crc kubenswrapper[4907]: I0313 16:12:48.543194 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" podStartSLOduration=10.052340781 podStartE2EDuration="10.543178325s" podCreationTimestamp="2026-03-13 16:12:38 +0000 UTC" firstStartedPulling="2026-03-13 16:12:39.331122958 +0000 UTC m=+7658.230910657" lastFinishedPulling="2026-03-13 16:12:39.821960512 +0000 UTC m=+7658.721748201" observedRunningTime="2026-03-13 16:12:40.469095794 +0000 UTC m=+7659.368883513" watchObservedRunningTime="2026-03-13 16:12:48.543178325 +0000 UTC m=+7667.442966014" Mar 13 16:12:56 crc kubenswrapper[4907]: I0313 16:12:56.633838 4907 generic.go:334] "Generic (PLEG): container finished" podID="cf92c1e7-1021-4405-840e-f2286adea31d" containerID="84b83daed0821729d3e57629749b6c8b5286ebc80f25b88784c7bd96792460bd" exitCode=0 Mar 13 16:12:56 crc kubenswrapper[4907]: I0313 16:12:56.633941 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerDied","Data":"84b83daed0821729d3e57629749b6c8b5286ebc80f25b88784c7bd96792460bd"} Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.196337 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286395 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286525 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286670 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.286702 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") pod \"cf92c1e7-1021-4405-840e-f2286adea31d\" (UID: \"cf92c1e7-1021-4405-840e-f2286adea31d\") " Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.293651 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6" (OuterVolumeSpecName: "kube-api-access-fsql6") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "kube-api-access-fsql6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.294779 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph" (OuterVolumeSpecName: "ceph") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.317104 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory" (OuterVolumeSpecName: "inventory") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.321829 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "cf92c1e7-1021-4405-840e-f2286adea31d" (UID: "cf92c1e7-1021-4405-840e-f2286adea31d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390251 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390472 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsql6\" (UniqueName: \"kubernetes.io/projected/cf92c1e7-1021-4405-840e-f2286adea31d-kube-api-access-fsql6\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390547 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.390621 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf92c1e7-1021-4405-840e-f2286adea31d-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.656189 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" event={"ID":"cf92c1e7-1021-4405-840e-f2286adea31d","Type":"ContainerDied","Data":"895e8f2c18c5de20a20d4cb067c8d1021996b85a83d2443bfe64b2080b187ce0"} Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.656439 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="895e8f2c18c5de20a20d4cb067c8d1021996b85a83d2443bfe64b2080b187ce0" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.656457 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x7hvj" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.753418 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-55pht"] Mar 13 16:12:58 crc kubenswrapper[4907]: E0313 16:12:58.754097 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf92c1e7-1021-4405-840e-f2286adea31d" containerName="reboot-os-openstack-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.754122 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf92c1e7-1021-4405-840e-f2286adea31d" containerName="reboot-os-openstack-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.754317 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf92c1e7-1021-4405-840e-f2286adea31d" containerName="reboot-os-openstack-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.755161 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.757417 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.757780 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.760902 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.763546 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.771358 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-55pht"] Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900084 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900130 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900174 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900278 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900306 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900399 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900462 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900483 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900514 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900542 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900595 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:58 crc kubenswrapper[4907]: I0313 16:12:58.900614 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002585 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002614 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002641 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002671 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002738 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002758 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002780 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002839 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002867 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.002941 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.007656 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.008697 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.009045 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.009548 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.009559 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.010559 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.011290 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.011628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.011659 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.012119 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.012488 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.021762 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"install-certs-openstack-openstack-cell1-55pht\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.071970 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:12:59 crc kubenswrapper[4907]: I0313 16:12:59.697514 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-55pht"] Mar 13 16:13:00 crc kubenswrapper[4907]: I0313 16:13:00.674300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerStarted","Data":"c3d0a8f0dcc500de771fd7594059297f23bc149de65101bf8d50df48f2d537c6"} Mar 13 16:13:01 crc kubenswrapper[4907]: I0313 16:13:01.683544 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerStarted","Data":"3281f97dcc70691f05829d00e2c76abe637c47e042b459a53a28abfa20ebaf42"} Mar 13 16:13:01 crc kubenswrapper[4907]: I0313 16:13:01.710098 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-55pht" podStartSLOduration=2.907388228 podStartE2EDuration="3.710077637s" podCreationTimestamp="2026-03-13 16:12:58 +0000 UTC" firstStartedPulling="2026-03-13 16:12:59.706589795 +0000 UTC m=+7678.606377484" lastFinishedPulling="2026-03-13 16:13:00.509279204 +0000 UTC m=+7679.409066893" observedRunningTime="2026-03-13 16:13:01.70580821 +0000 UTC m=+7680.605595909" watchObservedRunningTime="2026-03-13 16:13:01.710077637 +0000 UTC m=+7680.609865326" Mar 13 16:13:19 crc kubenswrapper[4907]: I0313 16:13:19.887085 4907 generic.go:334] "Generic (PLEG): container finished" podID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerID="3281f97dcc70691f05829d00e2c76abe637c47e042b459a53a28abfa20ebaf42" exitCode=0 Mar 13 16:13:19 crc kubenswrapper[4907]: I0313 16:13:19.887186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerDied","Data":"3281f97dcc70691f05829d00e2c76abe637c47e042b459a53a28abfa20ebaf42"} Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.456432 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486640 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486704 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486769 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486795 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486820 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.486892 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.492833 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.496311 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz" (OuterVolumeSpecName: "kube-api-access-qrltz") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "kube-api-access-qrltz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.502203 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.518035 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.520608 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.542658 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory" (OuterVolumeSpecName: "inventory") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.588649 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.588962 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589052 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589092 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589176 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") pod \"c32eadac-12ec-49b8-930e-bc6af99631e8\" (UID: \"c32eadac-12ec-49b8-930e-bc6af99631e8\") " Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589854 4907 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589873 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589899 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589910 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589919 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrltz\" (UniqueName: \"kubernetes.io/projected/c32eadac-12ec-49b8-930e-bc6af99631e8-kube-api-access-qrltz\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.589928 4907 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.592452 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.592661 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph" (OuterVolumeSpecName: "ceph") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.593121 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.593233 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.593859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.595141 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c32eadac-12ec-49b8-930e-bc6af99631e8" (UID: "c32eadac-12ec-49b8-930e-bc6af99631e8"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692241 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692307 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692321 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692333 4907 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692344 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.692353 4907 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c32eadac-12ec-49b8-930e-bc6af99631e8-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.909642 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-55pht" event={"ID":"c32eadac-12ec-49b8-930e-bc6af99631e8","Type":"ContainerDied","Data":"c3d0a8f0dcc500de771fd7594059297f23bc149de65101bf8d50df48f2d537c6"} Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.909688 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3d0a8f0dcc500de771fd7594059297f23bc149de65101bf8d50df48f2d537c6" Mar 13 16:13:21 crc kubenswrapper[4907]: I0313 16:13:21.909719 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-55pht" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.004061 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-tcvvp"] Mar 13 16:13:22 crc kubenswrapper[4907]: E0313 16:13:22.004715 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerName="install-certs-openstack-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.004740 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerName="install-certs-openstack-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.005066 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="c32eadac-12ec-49b8-930e-bc6af99631e8" containerName="install-certs-openstack-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.006080 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.010776 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.011474 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.012138 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.013820 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.021623 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-tcvvp"] Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201674 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201757 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201838 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.201946 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.304636 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.304845 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.305025 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.305158 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.309107 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.309142 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.309839 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.326099 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ceph-client-openstack-openstack-cell1-tcvvp\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.337389 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:22 crc kubenswrapper[4907]: I0313 16:13:22.926390 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-tcvvp"] Mar 13 16:13:23 crc kubenswrapper[4907]: I0313 16:13:23.941369 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerStarted","Data":"023c9098cc761ac6f84ee782fabb6a2460877d7ae1e899af2870592e5799025d"} Mar 13 16:13:23 crc kubenswrapper[4907]: I0313 16:13:23.941779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerStarted","Data":"2510c8b37338fdfac88a3c5515d0d7e6350abffa7916555d9078ef3f9701bbfd"} Mar 13 16:13:23 crc kubenswrapper[4907]: I0313 16:13:23.966045 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" podStartSLOduration=2.462028175 podStartE2EDuration="2.966025778s" podCreationTimestamp="2026-03-13 16:13:21 +0000 UTC" firstStartedPulling="2026-03-13 16:13:22.936953155 +0000 UTC m=+7701.836740844" lastFinishedPulling="2026-03-13 16:13:23.440950728 +0000 UTC m=+7702.340738447" observedRunningTime="2026-03-13 16:13:23.956669342 +0000 UTC m=+7702.856457071" watchObservedRunningTime="2026-03-13 16:13:23.966025778 +0000 UTC m=+7702.865813467" Mar 13 16:13:29 crc kubenswrapper[4907]: I0313 16:13:29.007442 4907 generic.go:334] "Generic (PLEG): container finished" podID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerID="023c9098cc761ac6f84ee782fabb6a2460877d7ae1e899af2870592e5799025d" exitCode=0 Mar 13 16:13:29 crc kubenswrapper[4907]: I0313 16:13:29.007546 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerDied","Data":"023c9098cc761ac6f84ee782fabb6a2460877d7ae1e899af2870592e5799025d"} Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.499672 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613741 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613869 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613954 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.613996 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") pod \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\" (UID: \"ef037fca-09ee-48ed-b2f7-fc65aea9155a\") " Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.627023 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph" (OuterVolumeSpecName: "ceph") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.627172 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj" (OuterVolumeSpecName: "kube-api-access-hvtsj") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "kube-api-access-hvtsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.647534 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory" (OuterVolumeSpecName: "inventory") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.658456 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "ef037fca-09ee-48ed-b2f7-fc65aea9155a" (UID: "ef037fca-09ee-48ed-b2f7-fc65aea9155a"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.717826 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.718149 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvtsj\" (UniqueName: \"kubernetes.io/projected/ef037fca-09ee-48ed-b2f7-fc65aea9155a-kube-api-access-hvtsj\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.718163 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:30 crc kubenswrapper[4907]: I0313 16:13:30.718204 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ef037fca-09ee-48ed-b2f7-fc65aea9155a-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.033161 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" event={"ID":"ef037fca-09ee-48ed-b2f7-fc65aea9155a","Type":"ContainerDied","Data":"2510c8b37338fdfac88a3c5515d0d7e6350abffa7916555d9078ef3f9701bbfd"} Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.033207 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2510c8b37338fdfac88a3c5515d0d7e6350abffa7916555d9078ef3f9701bbfd" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.033179 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-tcvvp" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.120026 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8dwgb"] Mar 13 16:13:31 crc kubenswrapper[4907]: E0313 16:13:31.121028 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerName="ceph-client-openstack-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.121105 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerName="ceph-client-openstack-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.121692 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef037fca-09ee-48ed-b2f7-fc65aea9155a" containerName="ceph-client-openstack-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.123565 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.126453 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.126811 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.126831 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.127052 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.127148 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.136950 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8dwgb"] Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228444 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228707 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.228811 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.229324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.331898 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.331972 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332092 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332122 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332175 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.332223 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.333138 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.336325 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.336427 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.338348 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.338963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.353592 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"ovn-openstack-openstack-cell1-8dwgb\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:31 crc kubenswrapper[4907]: I0313 16:13:31.498123 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:13:32 crc kubenswrapper[4907]: I0313 16:13:32.041072 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8dwgb"] Mar 13 16:13:33 crc kubenswrapper[4907]: I0313 16:13:33.058287 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerStarted","Data":"d1f808535be426e781215cf8fdab292d9eab6619345966de095be637979de839"} Mar 13 16:13:33 crc kubenswrapper[4907]: I0313 16:13:33.059057 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerStarted","Data":"3da1f54adfbf7e7555f40a0016c9c1e8f607c6f36a3249f1f9d7c62e57e1380d"} Mar 13 16:13:33 crc kubenswrapper[4907]: I0313 16:13:33.089730 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" podStartSLOduration=1.666269276 podStartE2EDuration="2.089704936s" podCreationTimestamp="2026-03-13 16:13:31 +0000 UTC" firstStartedPulling="2026-03-13 16:13:32.040355886 +0000 UTC m=+7710.940143595" lastFinishedPulling="2026-03-13 16:13:32.463791566 +0000 UTC m=+7711.363579255" observedRunningTime="2026-03-13 16:13:33.085037638 +0000 UTC m=+7711.984825337" watchObservedRunningTime="2026-03-13 16:13:33.089704936 +0000 UTC m=+7711.989492665" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.148539 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.150763 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.153873 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.154304 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.156574 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.158364 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.255849 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"auto-csr-approver-29556974-8sntm\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.366444 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"auto-csr-approver-29556974-8sntm\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.405834 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"auto-csr-approver-29556974-8sntm\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:00 crc kubenswrapper[4907]: I0313 16:14:00.481634 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:01 crc kubenswrapper[4907]: I0313 16:14:01.532161 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:14:02 crc kubenswrapper[4907]: I0313 16:14:02.455844 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556974-8sntm" event={"ID":"5d0046fb-05a0-47a0-912c-a9495e2a0835","Type":"ContainerStarted","Data":"a72b1649802cb4c03ecb12ae633e70af70234cd1f39e5f25f67919475783e4b7"} Mar 13 16:14:04 crc kubenswrapper[4907]: I0313 16:14:04.484747 4907 generic.go:334] "Generic (PLEG): container finished" podID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerID="048a0a595dbc6f4c8cb2226742f2beb5b639bec2395f93e0935301c145de8608" exitCode=0 Mar 13 16:14:04 crc kubenswrapper[4907]: I0313 16:14:04.484834 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556974-8sntm" event={"ID":"5d0046fb-05a0-47a0-912c-a9495e2a0835","Type":"ContainerDied","Data":"048a0a595dbc6f4c8cb2226742f2beb5b639bec2395f93e0935301c145de8608"} Mar 13 16:14:05 crc kubenswrapper[4907]: I0313 16:14:05.868674 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.009864 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") pod \"5d0046fb-05a0-47a0-912c-a9495e2a0835\" (UID: \"5d0046fb-05a0-47a0-912c-a9495e2a0835\") " Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.020823 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6" (OuterVolumeSpecName: "kube-api-access-q7kw6") pod "5d0046fb-05a0-47a0-912c-a9495e2a0835" (UID: "5d0046fb-05a0-47a0-912c-a9495e2a0835"). InnerVolumeSpecName "kube-api-access-q7kw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.113040 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7kw6\" (UniqueName: \"kubernetes.io/projected/5d0046fb-05a0-47a0-912c-a9495e2a0835-kube-api-access-q7kw6\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.508863 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556974-8sntm" event={"ID":"5d0046fb-05a0-47a0-912c-a9495e2a0835","Type":"ContainerDied","Data":"a72b1649802cb4c03ecb12ae633e70af70234cd1f39e5f25f67919475783e4b7"} Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.508932 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72b1649802cb4c03ecb12ae633e70af70234cd1f39e5f25f67919475783e4b7" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.508970 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556974-8sntm" Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.963460 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:14:06 crc kubenswrapper[4907]: I0313 16:14:06.972208 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556968-xbmhk"] Mar 13 16:14:07 crc kubenswrapper[4907]: I0313 16:14:07.795213 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfdfc46b-b215-491c-ab2f-2a108c2c0882" path="/var/lib/kubelet/pods/cfdfc46b-b215-491c-ab2f-2a108c2c0882/volumes" Mar 13 16:14:37 crc kubenswrapper[4907]: I0313 16:14:37.844771 4907 generic.go:334] "Generic (PLEG): container finished" podID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerID="d1f808535be426e781215cf8fdab292d9eab6619345966de095be637979de839" exitCode=0 Mar 13 16:14:37 crc kubenswrapper[4907]: I0313 16:14:37.844872 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerDied","Data":"d1f808535be426e781215cf8fdab292d9eab6619345966de095be637979de839"} Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.259990 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408812 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408873 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.408989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.409012 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.409119 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") pod \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\" (UID: \"47d984fd-dfa5-4a10-91c3-3c9a13edc91c\") " Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.414120 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph" (OuterVolumeSpecName: "ceph") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.420419 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv" (OuterVolumeSpecName: "kube-api-access-4zlnv") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "kube-api-access-4zlnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.432147 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.446041 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory" (OuterVolumeSpecName: "inventory") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.447497 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.455530 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "47d984fd-dfa5-4a10-91c3-3c9a13edc91c" (UID: "47d984fd-dfa5-4a10-91c3-3c9a13edc91c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511801 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zlnv\" (UniqueName: \"kubernetes.io/projected/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-kube-api-access-4zlnv\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511844 4907 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511861 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511874 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511905 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.511941 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/47d984fd-dfa5-4a10-91c3-3c9a13edc91c-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.870038 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" event={"ID":"47d984fd-dfa5-4a10-91c3-3c9a13edc91c","Type":"ContainerDied","Data":"3da1f54adfbf7e7555f40a0016c9c1e8f607c6f36a3249f1f9d7c62e57e1380d"} Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.870362 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8dwgb" Mar 13 16:14:39 crc kubenswrapper[4907]: I0313 16:14:39.870376 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3da1f54adfbf7e7555f40a0016c9c1e8f607c6f36a3249f1f9d7c62e57e1380d" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.002483 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-pwqj6"] Mar 13 16:14:40 crc kubenswrapper[4907]: E0313 16:14:40.003089 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerName="ovn-openstack-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003114 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerName="ovn-openstack-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: E0313 16:14:40.003152 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerName="oc" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003166 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerName="oc" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003526 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d984fd-dfa5-4a10-91c3-3c9a13edc91c" containerName="ovn-openstack-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.003564 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" containerName="oc" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.004871 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012489 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012682 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012794 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.012924 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.013021 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.013402 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.041273 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-pwqj6"] Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.126567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127113 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127233 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127407 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127614 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.127825 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.128005 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230417 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230696 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230800 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.230986 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.231154 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.231237 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.231353 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237478 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237499 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.237999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.239593 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.242956 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.249663 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"neutron-metadata-openstack-openstack-cell1-pwqj6\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.343526 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:14:40 crc kubenswrapper[4907]: I0313 16:14:40.950085 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-pwqj6"] Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.599064 4907 scope.go:117] "RemoveContainer" containerID="6f4aac0052e053a0aefd7363aac839de6a2ba5c8481dbf090c19d76f2294e95d" Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.905511 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerStarted","Data":"974721c911609f40407a7889cdfed0cb7412642260334ed010d57dbf5bb59c51"} Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.905927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerStarted","Data":"f4a4333655b463ae652caae8d8ffaa8b9979eafbf031d280da29757bdd65f7a4"} Mar 13 16:14:41 crc kubenswrapper[4907]: I0313 16:14:41.935555 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" podStartSLOduration=2.391647765 podStartE2EDuration="2.93550783s" podCreationTimestamp="2026-03-13 16:14:39 +0000 UTC" firstStartedPulling="2026-03-13 16:14:40.959663003 +0000 UTC m=+7779.859450732" lastFinishedPulling="2026-03-13 16:14:41.503523098 +0000 UTC m=+7780.403310797" observedRunningTime="2026-03-13 16:14:41.921312401 +0000 UTC m=+7780.821100100" watchObservedRunningTime="2026-03-13 16:14:41.93550783 +0000 UTC m=+7780.835295519" Mar 13 16:14:48 crc kubenswrapper[4907]: I0313 16:14:48.041288 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:14:48 crc kubenswrapper[4907]: I0313 16:14:48.041713 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.160391 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm"] Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.162194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.168063 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.168173 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.176250 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm"] Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.346191 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.346405 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.346587 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.449613 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.449813 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.450153 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.453127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.463333 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.481101 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"collect-profiles-29556975-rmwsm\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.494726 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:00 crc kubenswrapper[4907]: I0313 16:15:00.941576 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm"] Mar 13 16:15:00 crc kubenswrapper[4907]: W0313 16:15:00.943735 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1d3954c_686f_4c87_bebd_95a286df5781.slice/crio-7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3 WatchSource:0}: Error finding container 7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3: Status 404 returned error can't find the container with id 7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3 Mar 13 16:15:01 crc kubenswrapper[4907]: I0313 16:15:01.111890 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" event={"ID":"e1d3954c-686f-4c87-bebd-95a286df5781","Type":"ContainerStarted","Data":"7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3"} Mar 13 16:15:02 crc kubenswrapper[4907]: I0313 16:15:02.128651 4907 generic.go:334] "Generic (PLEG): container finished" podID="e1d3954c-686f-4c87-bebd-95a286df5781" containerID="5d7f3378e64b40f003c74c11f875b9d214af6de777aed6922c0bc466b32e0063" exitCode=0 Mar 13 16:15:02 crc kubenswrapper[4907]: I0313 16:15:02.128892 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" event={"ID":"e1d3954c-686f-4c87-bebd-95a286df5781","Type":"ContainerDied","Data":"5d7f3378e64b40f003c74c11f875b9d214af6de777aed6922c0bc466b32e0063"} Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.521807 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.714317 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") pod \"e1d3954c-686f-4c87-bebd-95a286df5781\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.714539 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") pod \"e1d3954c-686f-4c87-bebd-95a286df5781\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.714567 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") pod \"e1d3954c-686f-4c87-bebd-95a286df5781\" (UID: \"e1d3954c-686f-4c87-bebd-95a286df5781\") " Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.715049 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume" (OuterVolumeSpecName: "config-volume") pod "e1d3954c-686f-4c87-bebd-95a286df5781" (UID: "e1d3954c-686f-4c87-bebd-95a286df5781"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.716330 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1d3954c-686f-4c87-bebd-95a286df5781-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.720615 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e1d3954c-686f-4c87-bebd-95a286df5781" (UID: "e1d3954c-686f-4c87-bebd-95a286df5781"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.720779 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4" (OuterVolumeSpecName: "kube-api-access-cw7x4") pod "e1d3954c-686f-4c87-bebd-95a286df5781" (UID: "e1d3954c-686f-4c87-bebd-95a286df5781"). InnerVolumeSpecName "kube-api-access-cw7x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.817385 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1d3954c-686f-4c87-bebd-95a286df5781-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:03 crc kubenswrapper[4907]: I0313 16:15:03.817725 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw7x4\" (UniqueName: \"kubernetes.io/projected/e1d3954c-686f-4c87-bebd-95a286df5781-kube-api-access-cw7x4\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.154002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" event={"ID":"e1d3954c-686f-4c87-bebd-95a286df5781","Type":"ContainerDied","Data":"7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3"} Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.154054 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556975-rmwsm" Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.154061 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7110db02ed258c363241254fa3b18e476c70ce370c9163d88c74a6f6dca09ec3" Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.602077 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 16:15:04 crc kubenswrapper[4907]: I0313 16:15:04.611643 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556930-f852z"] Mar 13 16:15:05 crc kubenswrapper[4907]: I0313 16:15:05.794788 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2983956e-5d8d-4595-9fda-91759741010a" path="/var/lib/kubelet/pods/2983956e-5d8d-4595-9fda-91759741010a/volumes" Mar 13 16:15:18 crc kubenswrapper[4907]: I0313 16:15:18.042201 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:15:18 crc kubenswrapper[4907]: I0313 16:15:18.042723 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:15:34 crc kubenswrapper[4907]: I0313 16:15:34.464470 4907 generic.go:334] "Generic (PLEG): container finished" podID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerID="974721c911609f40407a7889cdfed0cb7412642260334ed010d57dbf5bb59c51" exitCode=0 Mar 13 16:15:34 crc kubenswrapper[4907]: I0313 16:15:34.464556 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerDied","Data":"974721c911609f40407a7889cdfed0cb7412642260334ed010d57dbf5bb59c51"} Mar 13 16:15:35 crc kubenswrapper[4907]: I0313 16:15:35.955103 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.135978 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136214 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136272 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136549 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136601 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136700 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.136758 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") pod \"3dc18cfe-d63a-4960-8c7d-10913cc78896\" (UID: \"3dc18cfe-d63a-4960-8c7d-10913cc78896\") " Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.145481 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6" (OuterVolumeSpecName: "kube-api-access-bcsp6") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "kube-api-access-bcsp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.146231 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph" (OuterVolumeSpecName: "ceph") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.147896 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.172368 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.174183 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.174556 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.175981 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory" (OuterVolumeSpecName: "inventory") pod "3dc18cfe-d63a-4960-8c7d-10913cc78896" (UID: "3dc18cfe-d63a-4960-8c7d-10913cc78896"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241751 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241809 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241822 4907 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241864 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241874 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241903 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcsp6\" (UniqueName: \"kubernetes.io/projected/3dc18cfe-d63a-4960-8c7d-10913cc78896-kube-api-access-bcsp6\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.241915 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dc18cfe-d63a-4960-8c7d-10913cc78896-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.485805 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" event={"ID":"3dc18cfe-d63a-4960-8c7d-10913cc78896","Type":"ContainerDied","Data":"f4a4333655b463ae652caae8d8ffaa8b9979eafbf031d280da29757bdd65f7a4"} Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.486215 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4a4333655b463ae652caae8d8ffaa8b9979eafbf031d280da29757bdd65f7a4" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.485861 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-pwqj6" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.654328 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-bdwbh"] Mar 13 16:15:36 crc kubenswrapper[4907]: E0313 16:15:36.654800 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d3954c-686f-4c87-bebd-95a286df5781" containerName="collect-profiles" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.654821 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d3954c-686f-4c87-bebd-95a286df5781" containerName="collect-profiles" Mar 13 16:15:36 crc kubenswrapper[4907]: E0313 16:15:36.654843 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerName="neutron-metadata-openstack-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.654854 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerName="neutron-metadata-openstack-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.655155 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d3954c-686f-4c87-bebd-95a286df5781" containerName="collect-profiles" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.655176 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc18cfe-d63a-4960-8c7d-10913cc78896" containerName="neutron-metadata-openstack-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.656017 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.660137 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.660285 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.660653 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.661351 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.661508 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.676518 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-bdwbh"] Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751146 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751190 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751221 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751383 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751575 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.751703 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853517 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853598 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853647 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853699 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.853828 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.857863 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.858339 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.859650 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.860356 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.862771 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:36 crc kubenswrapper[4907]: I0313 16:15:36.871534 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"libvirt-openstack-openstack-cell1-bdwbh\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:37 crc kubenswrapper[4907]: I0313 16:15:37.036057 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:15:37 crc kubenswrapper[4907]: I0313 16:15:37.588160 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-bdwbh"] Mar 13 16:15:38 crc kubenswrapper[4907]: I0313 16:15:38.505874 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerStarted","Data":"9ec4cbc3bc8367d5cb8d52b834231c90815a53ebf59eb9499587563be45dd71a"} Mar 13 16:15:38 crc kubenswrapper[4907]: I0313 16:15:38.506562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerStarted","Data":"e4c7c890b0bd5338464cbbe531bda4c9b8efa360ef93c0753d3a22a75540214d"} Mar 13 16:15:38 crc kubenswrapper[4907]: I0313 16:15:38.532839 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" podStartSLOduration=2.051412183 podStartE2EDuration="2.532821719s" podCreationTimestamp="2026-03-13 16:15:36 +0000 UTC" firstStartedPulling="2026-03-13 16:15:37.597014747 +0000 UTC m=+7836.496802436" lastFinishedPulling="2026-03-13 16:15:38.078424283 +0000 UTC m=+7836.978211972" observedRunningTime="2026-03-13 16:15:38.520227314 +0000 UTC m=+7837.420015013" watchObservedRunningTime="2026-03-13 16:15:38.532821719 +0000 UTC m=+7837.432609408" Mar 13 16:15:41 crc kubenswrapper[4907]: I0313 16:15:41.775721 4907 scope.go:117] "RemoveContainer" containerID="74d889f05e8d8fa3b0cfc7fed167cff28426e276bf0f0c9f759dd1496f3c1058" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.042308 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.042966 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.043015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.043913 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.043980 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" gracePeriod=600 Mar 13 16:15:48 crc kubenswrapper[4907]: E0313 16:15:48.198077 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.661946 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" exitCode=0 Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.661998 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f"} Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.662035 4907 scope.go:117] "RemoveContainer" containerID="c38b8b19c660dee13d11b404db7a1f6a1a7c1615cf40ea14a2875b4e3648d560" Mar 13 16:15:48 crc kubenswrapper[4907]: I0313 16:15:48.663965 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:15:48 crc kubenswrapper[4907]: E0313 16:15:48.664397 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.148576 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.151113 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.154660 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.155178 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.160064 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.162933 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.257310 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"auto-csr-approver-29556976-twcx4\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.359257 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"auto-csr-approver-29556976-twcx4\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.378200 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"auto-csr-approver-29556976-twcx4\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.473491 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:00 crc kubenswrapper[4907]: I0313 16:16:00.941709 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:16:00 crc kubenswrapper[4907]: W0313 16:16:00.950064 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod362f626f_a4dd_45a7_ad88_4c539732eebb.slice/crio-bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe WatchSource:0}: Error finding container bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe: Status 404 returned error can't find the container with id bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe Mar 13 16:16:01 crc kubenswrapper[4907]: I0313 16:16:01.802127 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556976-twcx4" event={"ID":"362f626f-a4dd-45a7-ad88-4c539732eebb","Type":"ContainerStarted","Data":"bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe"} Mar 13 16:16:01 crc kubenswrapper[4907]: I0313 16:16:01.802636 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:01 crc kubenswrapper[4907]: E0313 16:16:01.803158 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:02 crc kubenswrapper[4907]: I0313 16:16:02.818020 4907 generic.go:334] "Generic (PLEG): container finished" podID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerID="7e0af7bb09c83c72fd3ad1550fb6bc5bec78d9ec9c71037ff20efe437fc967c9" exitCode=0 Mar 13 16:16:02 crc kubenswrapper[4907]: I0313 16:16:02.818119 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556976-twcx4" event={"ID":"362f626f-a4dd-45a7-ad88-4c539732eebb","Type":"ContainerDied","Data":"7e0af7bb09c83c72fd3ad1550fb6bc5bec78d9ec9c71037ff20efe437fc967c9"} Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.192694 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.345014 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") pod \"362f626f-a4dd-45a7-ad88-4c539732eebb\" (UID: \"362f626f-a4dd-45a7-ad88-4c539732eebb\") " Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.350467 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw" (OuterVolumeSpecName: "kube-api-access-5vkvw") pod "362f626f-a4dd-45a7-ad88-4c539732eebb" (UID: "362f626f-a4dd-45a7-ad88-4c539732eebb"). InnerVolumeSpecName "kube-api-access-5vkvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.450579 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vkvw\" (UniqueName: \"kubernetes.io/projected/362f626f-a4dd-45a7-ad88-4c539732eebb-kube-api-access-5vkvw\") on node \"crc\" DevicePath \"\"" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.857349 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556976-twcx4" event={"ID":"362f626f-a4dd-45a7-ad88-4c539732eebb","Type":"ContainerDied","Data":"bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe"} Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.857976 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bda227a271f38ea0cca40341cba0484c8a2a657702d708161303a4775090dbbe" Mar 13 16:16:04 crc kubenswrapper[4907]: I0313 16:16:04.857413 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556976-twcx4" Mar 13 16:16:05 crc kubenswrapper[4907]: I0313 16:16:05.291125 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:16:05 crc kubenswrapper[4907]: I0313 16:16:05.302838 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556970-lbxp9"] Mar 13 16:16:05 crc kubenswrapper[4907]: I0313 16:16:05.796231 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b3a466a-0c1c-4c86-897f-0a3375a9fb53" path="/var/lib/kubelet/pods/2b3a466a-0c1c-4c86-897f-0a3375a9fb53/volumes" Mar 13 16:16:14 crc kubenswrapper[4907]: I0313 16:16:14.782824 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:14 crc kubenswrapper[4907]: E0313 16:16:14.783786 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:29 crc kubenswrapper[4907]: I0313 16:16:29.782966 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:29 crc kubenswrapper[4907]: E0313 16:16:29.784070 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:41 crc kubenswrapper[4907]: I0313 16:16:41.862222 4907 scope.go:117] "RemoveContainer" containerID="1ffb094832567d23bf593fb1872edfb355e12059011cdaa6ddadd04821d723cc" Mar 13 16:16:44 crc kubenswrapper[4907]: I0313 16:16:44.783282 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:44 crc kubenswrapper[4907]: E0313 16:16:44.784329 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:16:59 crc kubenswrapper[4907]: I0313 16:16:59.782737 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:16:59 crc kubenswrapper[4907]: E0313 16:16:59.785234 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:14 crc kubenswrapper[4907]: I0313 16:17:14.781978 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:14 crc kubenswrapper[4907]: E0313 16:17:14.782788 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:29 crc kubenswrapper[4907]: I0313 16:17:29.784023 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:29 crc kubenswrapper[4907]: E0313 16:17:29.787066 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:40 crc kubenswrapper[4907]: I0313 16:17:40.782697 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:40 crc kubenswrapper[4907]: E0313 16:17:40.783535 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.787184 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:17:48 crc kubenswrapper[4907]: E0313 16:17:48.788581 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerName="oc" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.788602 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerName="oc" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.788912 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" containerName="oc" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.791004 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.824136 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.870702 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.870803 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.871309 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.973909 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.973998 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.974133 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.974680 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:48 crc kubenswrapper[4907]: I0313 16:17:48.975001 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:49 crc kubenswrapper[4907]: I0313 16:17:49.006958 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"redhat-marketplace-4cppb\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:49 crc kubenswrapper[4907]: I0313 16:17:49.131786 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:49 crc kubenswrapper[4907]: I0313 16:17:49.598564 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.080461 4907 generic.go:334] "Generic (PLEG): container finished" podID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" exitCode=0 Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.080512 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03"} Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.080563 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerStarted","Data":"5a312c55abb195fc381c69253b66e281eba78efb1b39b9a51a72abfc897857dc"} Mar 13 16:17:50 crc kubenswrapper[4907]: I0313 16:17:50.085521 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:17:51 crc kubenswrapper[4907]: I0313 16:17:51.091325 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerStarted","Data":"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f"} Mar 13 16:17:52 crc kubenswrapper[4907]: I0313 16:17:52.105096 4907 generic.go:334] "Generic (PLEG): container finished" podID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" exitCode=0 Mar 13 16:17:52 crc kubenswrapper[4907]: I0313 16:17:52.105150 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f"} Mar 13 16:17:53 crc kubenswrapper[4907]: I0313 16:17:53.116116 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerStarted","Data":"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6"} Mar 13 16:17:53 crc kubenswrapper[4907]: I0313 16:17:53.137262 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4cppb" podStartSLOduration=2.604871424 podStartE2EDuration="5.137244139s" podCreationTimestamp="2026-03-13 16:17:48 +0000 UTC" firstStartedPulling="2026-03-13 16:17:50.084665006 +0000 UTC m=+7968.984452705" lastFinishedPulling="2026-03-13 16:17:52.617037731 +0000 UTC m=+7971.516825420" observedRunningTime="2026-03-13 16:17:53.132621932 +0000 UTC m=+7972.032409621" watchObservedRunningTime="2026-03-13 16:17:53.137244139 +0000 UTC m=+7972.037031828" Mar 13 16:17:55 crc kubenswrapper[4907]: I0313 16:17:55.782616 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:17:55 crc kubenswrapper[4907]: E0313 16:17:55.783467 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.132637 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.134092 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.189682 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.260862 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:17:59 crc kubenswrapper[4907]: I0313 16:17:59.442679 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.154332 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.157192 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.160813 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.161167 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.161414 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.167205 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.225477 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"auto-csr-approver-29556978-m5zzz\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.327835 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"auto-csr-approver-29556978-m5zzz\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.360212 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"auto-csr-approver-29556978-m5zzz\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.486707 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:00 crc kubenswrapper[4907]: I0313 16:18:00.977999 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.203498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" event={"ID":"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1","Type":"ContainerStarted","Data":"432517852f22b2935c9497d6a30eeb142c617505442bf2e407018010babc7523"} Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.203655 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4cppb" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" containerID="cri-o://5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" gracePeriod=2 Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.714804 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.885028 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") pod \"d417de61-9b72-4c15-bfae-6f1b6ee03849\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.885114 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") pod \"d417de61-9b72-4c15-bfae-6f1b6ee03849\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.885362 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") pod \"d417de61-9b72-4c15-bfae-6f1b6ee03849\" (UID: \"d417de61-9b72-4c15-bfae-6f1b6ee03849\") " Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.886963 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities" (OuterVolumeSpecName: "utilities") pod "d417de61-9b72-4c15-bfae-6f1b6ee03849" (UID: "d417de61-9b72-4c15-bfae-6f1b6ee03849"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.892765 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk" (OuterVolumeSpecName: "kube-api-access-vf7bk") pod "d417de61-9b72-4c15-bfae-6f1b6ee03849" (UID: "d417de61-9b72-4c15-bfae-6f1b6ee03849"). InnerVolumeSpecName "kube-api-access-vf7bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.926851 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d417de61-9b72-4c15-bfae-6f1b6ee03849" (UID: "d417de61-9b72-4c15-bfae-6f1b6ee03849"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.987628 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.987672 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d417de61-9b72-4c15-bfae-6f1b6ee03849-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:01 crc kubenswrapper[4907]: I0313 16:18:01.987683 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf7bk\" (UniqueName: \"kubernetes.io/projected/d417de61-9b72-4c15-bfae-6f1b6ee03849-kube-api-access-vf7bk\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.213665 4907 generic.go:334] "Generic (PLEG): container finished" podID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" exitCode=0 Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214030 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6"} Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214059 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4cppb" event={"ID":"d417de61-9b72-4c15-bfae-6f1b6ee03849","Type":"ContainerDied","Data":"5a312c55abb195fc381c69253b66e281eba78efb1b39b9a51a72abfc897857dc"} Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214077 4907 scope.go:117] "RemoveContainer" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.214219 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4cppb" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.242894 4907 scope.go:117] "RemoveContainer" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.256554 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.277375 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4cppb"] Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.281476 4907 scope.go:117] "RemoveContainer" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.325168 4907 scope.go:117] "RemoveContainer" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" Mar 13 16:18:02 crc kubenswrapper[4907]: E0313 16:18:02.325838 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6\": container with ID starting with 5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6 not found: ID does not exist" containerID="5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.325973 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6"} err="failed to get container status \"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6\": rpc error: code = NotFound desc = could not find container \"5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6\": container with ID starting with 5bcc136eb40ce02d1e8aec26066c465f5d2b04e99adae87f346f7d3227130ae6 not found: ID does not exist" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.326018 4907 scope.go:117] "RemoveContainer" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" Mar 13 16:18:02 crc kubenswrapper[4907]: E0313 16:18:02.327030 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f\": container with ID starting with cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f not found: ID does not exist" containerID="cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.327068 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f"} err="failed to get container status \"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f\": rpc error: code = NotFound desc = could not find container \"cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f\": container with ID starting with cc8ae91cc5b08e53b842e2f7a69a2fbd79580167c23866df62a867359907d98f not found: ID does not exist" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.327089 4907 scope.go:117] "RemoveContainer" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" Mar 13 16:18:02 crc kubenswrapper[4907]: E0313 16:18:02.327331 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03\": container with ID starting with e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03 not found: ID does not exist" containerID="e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03" Mar 13 16:18:02 crc kubenswrapper[4907]: I0313 16:18:02.327362 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03"} err="failed to get container status \"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03\": rpc error: code = NotFound desc = could not find container \"e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03\": container with ID starting with e3c0406c3eaa17bc2c3ae343da02515c95a6e17e38c2bc85e8e82bdf569d1c03 not found: ID does not exist" Mar 13 16:18:03 crc kubenswrapper[4907]: I0313 16:18:03.231461 4907 generic.go:334] "Generic (PLEG): container finished" podID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerID="f49981cfe3caeb1a9f80b1991ed24d10adb510db888dce88031b1e2cc8045b1b" exitCode=0 Mar 13 16:18:03 crc kubenswrapper[4907]: I0313 16:18:03.231765 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" event={"ID":"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1","Type":"ContainerDied","Data":"f49981cfe3caeb1a9f80b1991ed24d10adb510db888dce88031b1e2cc8045b1b"} Mar 13 16:18:03 crc kubenswrapper[4907]: I0313 16:18:03.798729 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" path="/var/lib/kubelet/pods/d417de61-9b72-4c15-bfae-6f1b6ee03849/volumes" Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.694515 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.860686 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") pod \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\" (UID: \"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1\") " Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.867289 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr" (OuterVolumeSpecName: "kube-api-access-rhvqr") pod "b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" (UID: "b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1"). InnerVolumeSpecName "kube-api-access-rhvqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:18:04 crc kubenswrapper[4907]: I0313 16:18:04.963711 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhvqr\" (UniqueName: \"kubernetes.io/projected/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1-kube-api-access-rhvqr\") on node \"crc\" DevicePath \"\"" Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.255954 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" event={"ID":"b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1","Type":"ContainerDied","Data":"432517852f22b2935c9497d6a30eeb142c617505442bf2e407018010babc7523"} Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.255998 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="432517852f22b2935c9497d6a30eeb142c617505442bf2e407018010babc7523" Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.256025 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556978-m5zzz" Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.755397 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.763980 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556972-llpxn"] Mar 13 16:18:05 crc kubenswrapper[4907]: I0313 16:18:05.794331 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85500c41-93c1-44ae-b418-d7f125c04621" path="/var/lib/kubelet/pods/85500c41-93c1-44ae-b418-d7f125c04621/volumes" Mar 13 16:18:07 crc kubenswrapper[4907]: I0313 16:18:07.785912 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:07 crc kubenswrapper[4907]: E0313 16:18:07.786578 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:21 crc kubenswrapper[4907]: I0313 16:18:21.792937 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:21 crc kubenswrapper[4907]: E0313 16:18:21.794101 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:32 crc kubenswrapper[4907]: I0313 16:18:32.782583 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:32 crc kubenswrapper[4907]: E0313 16:18:32.783395 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:42 crc kubenswrapper[4907]: I0313 16:18:42.035904 4907 scope.go:117] "RemoveContainer" containerID="1597ac303cc302238cff89d1286c78c824124bc23ca555a00a5dd03f41ac4104" Mar 13 16:18:43 crc kubenswrapper[4907]: I0313 16:18:43.784988 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:43 crc kubenswrapper[4907]: E0313 16:18:43.786003 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:18:55 crc kubenswrapper[4907]: I0313 16:18:55.784044 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:18:55 crc kubenswrapper[4907]: E0313 16:18:55.784876 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:08 crc kubenswrapper[4907]: I0313 16:19:08.782671 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:08 crc kubenswrapper[4907]: E0313 16:19:08.783476 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:22 crc kubenswrapper[4907]: I0313 16:19:22.783422 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:22 crc kubenswrapper[4907]: E0313 16:19:22.784404 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:36 crc kubenswrapper[4907]: I0313 16:19:36.783137 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:36 crc kubenswrapper[4907]: E0313 16:19:36.784158 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:19:50 crc kubenswrapper[4907]: I0313 16:19:50.782669 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:19:50 crc kubenswrapper[4907]: E0313 16:19:50.783596 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.156934 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157872 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-utilities" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157905 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-utilities" Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157913 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157921 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157952 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-content" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157959 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="extract-content" Mar 13 16:20:00 crc kubenswrapper[4907]: E0313 16:20:00.157967 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerName="oc" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.157973 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerName="oc" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.158221 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d417de61-9b72-4c15-bfae-6f1b6ee03849" containerName="registry-server" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.158237 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" containerName="oc" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.158979 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.162324 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.169270 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.170994 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.172134 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.226148 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"auto-csr-approver-29556980-5b2q4\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.328704 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"auto-csr-approver-29556980-5b2q4\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.348644 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"auto-csr-approver-29556980-5b2q4\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:00 crc kubenswrapper[4907]: I0313 16:20:00.493090 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:01 crc kubenswrapper[4907]: W0313 16:20:01.045798 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b9615a4_0270_4660_bd34_5efded647a80.slice/crio-1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601 WatchSource:0}: Error finding container 1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601: Status 404 returned error can't find the container with id 1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601 Mar 13 16:20:01 crc kubenswrapper[4907]: I0313 16:20:01.049257 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:20:01 crc kubenswrapper[4907]: I0313 16:20:01.642504 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" event={"ID":"0b9615a4-0270-4660-bd34-5efded647a80","Type":"ContainerStarted","Data":"1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601"} Mar 13 16:20:03 crc kubenswrapper[4907]: I0313 16:20:03.674521 4907 generic.go:334] "Generic (PLEG): container finished" podID="0b9615a4-0270-4660-bd34-5efded647a80" containerID="d626b98830070c30af12d3ea7ba374e64df5c680e5d7d98055b77c7c0b8bf280" exitCode=0 Mar 13 16:20:03 crc kubenswrapper[4907]: I0313 16:20:03.674640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" event={"ID":"0b9615a4-0270-4660-bd34-5efded647a80","Type":"ContainerDied","Data":"d626b98830070c30af12d3ea7ba374e64df5c680e5d7d98055b77c7c0b8bf280"} Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.043240 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.149446 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") pod \"0b9615a4-0270-4660-bd34-5efded647a80\" (UID: \"0b9615a4-0270-4660-bd34-5efded647a80\") " Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.156473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6" (OuterVolumeSpecName: "kube-api-access-cg2d6") pod "0b9615a4-0270-4660-bd34-5efded647a80" (UID: "0b9615a4-0270-4660-bd34-5efded647a80"). InnerVolumeSpecName "kube-api-access-cg2d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.253607 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg2d6\" (UniqueName: \"kubernetes.io/projected/0b9615a4-0270-4660-bd34-5efded647a80-kube-api-access-cg2d6\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.707641 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" event={"ID":"0b9615a4-0270-4660-bd34-5efded647a80","Type":"ContainerDied","Data":"1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601"} Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.707710 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0bbe7efbc6537e7c285a15ed6f9873236deb0bf8e950cf4296237e90a80601" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.707801 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556980-5b2q4" Mar 13 16:20:05 crc kubenswrapper[4907]: I0313 16:20:05.785227 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:05 crc kubenswrapper[4907]: E0313 16:20:05.785636 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:06 crc kubenswrapper[4907]: I0313 16:20:06.135966 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:20:06 crc kubenswrapper[4907]: I0313 16:20:06.149596 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556974-8sntm"] Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.561590 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:07 crc kubenswrapper[4907]: E0313 16:20:07.562469 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b9615a4-0270-4660-bd34-5efded647a80" containerName="oc" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.562487 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b9615a4-0270-4660-bd34-5efded647a80" containerName="oc" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.562786 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b9615a4-0270-4660-bd34-5efded647a80" containerName="oc" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.564872 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.580785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.719863 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.720055 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.720587 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.728598 4907 generic.go:334] "Generic (PLEG): container finished" podID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerID="9ec4cbc3bc8367d5cb8d52b834231c90815a53ebf59eb9499587563be45dd71a" exitCode=0 Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.728640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerDied","Data":"9ec4cbc3bc8367d5cb8d52b834231c90815a53ebf59eb9499587563be45dd71a"} Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.795459 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d0046fb-05a0-47a0-912c-a9495e2a0835" path="/var/lib/kubelet/pods/5d0046fb-05a0-47a0-912c-a9495e2a0835/volumes" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822320 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822477 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822984 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.822999 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.851284 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"redhat-operators-p5chz\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:07 crc kubenswrapper[4907]: I0313 16:20:07.898270 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.394044 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.754807 4907 generic.go:334] "Generic (PLEG): container finished" podID="e879dfd4-a757-496a-9255-3905537d4c2f" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" exitCode=0 Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.755247 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85"} Mar 13 16:20:08 crc kubenswrapper[4907]: I0313 16:20:08.755308 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerStarted","Data":"b255de30c32bfb6d5b4d9175c23c14378544689fd3ec3a2b3dc3a46fb608cce3"} Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.197979 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369142 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369278 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369319 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369391 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.369689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.370274 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") pod \"82c7c5cb-4f6b-4d38-a009-263025faa45b\" (UID: \"82c7c5cb-4f6b-4d38-a009-263025faa45b\") " Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.374575 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph" (OuterVolumeSpecName: "ceph") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.375652 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.375753 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5" (OuterVolumeSpecName: "kube-api-access-h7hx5") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "kube-api-access-h7hx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.398656 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.399302 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.400413 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory" (OuterVolumeSpecName: "inventory") pod "82c7c5cb-4f6b-4d38-a009-263025faa45b" (UID: "82c7c5cb-4f6b-4d38-a009-263025faa45b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473134 4907 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473174 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7hx5\" (UniqueName: \"kubernetes.io/projected/82c7c5cb-4f6b-4d38-a009-263025faa45b-kube-api-access-h7hx5\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473188 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473197 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473206 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.473215 4907 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82c7c5cb-4f6b-4d38-a009-263025faa45b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.775260 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" event={"ID":"82c7c5cb-4f6b-4d38-a009-263025faa45b","Type":"ContainerDied","Data":"e4c7c890b0bd5338464cbbe531bda4c9b8efa360ef93c0753d3a22a75540214d"} Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.775355 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4c7c890b0bd5338464cbbe531bda4c9b8efa360ef93c0753d3a22a75540214d" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.775324 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-bdwbh" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.858782 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-2swxs"] Mar 13 16:20:09 crc kubenswrapper[4907]: E0313 16:20:09.859699 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerName="libvirt-openstack-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.859834 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerName="libvirt-openstack-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.860408 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c7c5cb-4f6b-4d38-a009-263025faa45b" containerName="libvirt-openstack-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.862194 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.864852 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865070 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865306 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865597 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865815 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865647 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.865662 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.898664 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-2swxs"] Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.982902 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.982968 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983038 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983079 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983130 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983156 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983178 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983210 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983235 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983282 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983303 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:09 crc kubenswrapper[4907]: I0313 16:20:09.983351 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084745 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084795 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084866 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084960 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.084995 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085024 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085063 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085100 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085123 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085174 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.085195 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.086229 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.087667 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.089071 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.090543 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.090655 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.090742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.091085 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.091479 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.091642 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.092128 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.093584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.098176 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.113597 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"nova-cell1-openstack-openstack-cell1-2swxs\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.183066 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.718388 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-2swxs"] Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.786548 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerStarted","Data":"f3963807c8d98e582dff2a1a7b85c9b8817c6b50cae040491cc0e2d933ebfd1c"} Mar 13 16:20:10 crc kubenswrapper[4907]: I0313 16:20:10.789305 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerStarted","Data":"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594"} Mar 13 16:20:11 crc kubenswrapper[4907]: I0313 16:20:11.802108 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerStarted","Data":"24ef5c64c1022adaf88de5eb0ff8963c4a40b5475aaa5f2ebc31c41febbdb0f6"} Mar 13 16:20:11 crc kubenswrapper[4907]: I0313 16:20:11.840419 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" podStartSLOduration=2.216571227 podStartE2EDuration="2.84039971s" podCreationTimestamp="2026-03-13 16:20:09 +0000 UTC" firstStartedPulling="2026-03-13 16:20:10.72265324 +0000 UTC m=+8109.622440929" lastFinishedPulling="2026-03-13 16:20:11.346481723 +0000 UTC m=+8110.246269412" observedRunningTime="2026-03-13 16:20:11.83855202 +0000 UTC m=+8110.738339729" watchObservedRunningTime="2026-03-13 16:20:11.84039971 +0000 UTC m=+8110.740187409" Mar 13 16:20:14 crc kubenswrapper[4907]: I0313 16:20:14.835848 4907 generic.go:334] "Generic (PLEG): container finished" podID="e879dfd4-a757-496a-9255-3905537d4c2f" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" exitCode=0 Mar 13 16:20:14 crc kubenswrapper[4907]: I0313 16:20:14.835928 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594"} Mar 13 16:20:15 crc kubenswrapper[4907]: I0313 16:20:15.848586 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerStarted","Data":"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e"} Mar 13 16:20:15 crc kubenswrapper[4907]: I0313 16:20:15.869510 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p5chz" podStartSLOduration=2.294996328 podStartE2EDuration="8.869494769s" podCreationTimestamp="2026-03-13 16:20:07 +0000 UTC" firstStartedPulling="2026-03-13 16:20:08.761747144 +0000 UTC m=+8107.661534833" lastFinishedPulling="2026-03-13 16:20:15.336245595 +0000 UTC m=+8114.236033274" observedRunningTime="2026-03-13 16:20:15.867219887 +0000 UTC m=+8114.767007576" watchObservedRunningTime="2026-03-13 16:20:15.869494769 +0000 UTC m=+8114.769282458" Mar 13 16:20:16 crc kubenswrapper[4907]: I0313 16:20:16.782029 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:16 crc kubenswrapper[4907]: E0313 16:20:16.782348 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:17 crc kubenswrapper[4907]: I0313 16:20:17.899528 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:17 crc kubenswrapper[4907]: I0313 16:20:17.899988 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:18 crc kubenswrapper[4907]: I0313 16:20:18.955588 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p5chz" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" probeResult="failure" output=< Mar 13 16:20:18 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:20:18 crc kubenswrapper[4907]: > Mar 13 16:20:27 crc kubenswrapper[4907]: I0313 16:20:27.965705 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:28 crc kubenswrapper[4907]: I0313 16:20:28.041491 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:28 crc kubenswrapper[4907]: I0313 16:20:28.216414 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.041271 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p5chz" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" containerID="cri-o://77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" gracePeriod=2 Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.567228 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.641976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") pod \"e879dfd4-a757-496a-9255-3905537d4c2f\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.642245 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") pod \"e879dfd4-a757-496a-9255-3905537d4c2f\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.642519 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") pod \"e879dfd4-a757-496a-9255-3905537d4c2f\" (UID: \"e879dfd4-a757-496a-9255-3905537d4c2f\") " Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.643290 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities" (OuterVolumeSpecName: "utilities") pod "e879dfd4-a757-496a-9255-3905537d4c2f" (UID: "e879dfd4-a757-496a-9255-3905537d4c2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.647731 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99" (OuterVolumeSpecName: "kube-api-access-48v99") pod "e879dfd4-a757-496a-9255-3905537d4c2f" (UID: "e879dfd4-a757-496a-9255-3905537d4c2f"). InnerVolumeSpecName "kube-api-access-48v99". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.744768 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.744796 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48v99\" (UniqueName: \"kubernetes.io/projected/e879dfd4-a757-496a-9255-3905537d4c2f-kube-api-access-48v99\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.793132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e879dfd4-a757-496a-9255-3905537d4c2f" (UID: "e879dfd4-a757-496a-9255-3905537d4c2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:20:29 crc kubenswrapper[4907]: I0313 16:20:29.848170 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e879dfd4-a757-496a-9255-3905537d4c2f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059305 4907 generic.go:334] "Generic (PLEG): container finished" podID="e879dfd4-a757-496a-9255-3905537d4c2f" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" exitCode=0 Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059338 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5chz" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059397 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e"} Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059480 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5chz" event={"ID":"e879dfd4-a757-496a-9255-3905537d4c2f","Type":"ContainerDied","Data":"b255de30c32bfb6d5b4d9175c23c14378544689fd3ec3a2b3dc3a46fb608cce3"} Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.059516 4907 scope.go:117] "RemoveContainer" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.090752 4907 scope.go:117] "RemoveContainer" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.098783 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.112589 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p5chz"] Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.121671 4907 scope.go:117] "RemoveContainer" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.190099 4907 scope.go:117] "RemoveContainer" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" Mar 13 16:20:30 crc kubenswrapper[4907]: E0313 16:20:30.190653 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e\": container with ID starting with 77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e not found: ID does not exist" containerID="77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.190729 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e"} err="failed to get container status \"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e\": rpc error: code = NotFound desc = could not find container \"77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e\": container with ID starting with 77ec1446661bc5f367d029f3cae8705d509f9b1e3cacfc9bceeb4dc3c70ed20e not found: ID does not exist" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.190772 4907 scope.go:117] "RemoveContainer" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" Mar 13 16:20:30 crc kubenswrapper[4907]: E0313 16:20:30.191368 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594\": container with ID starting with 35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594 not found: ID does not exist" containerID="35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.191409 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594"} err="failed to get container status \"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594\": rpc error: code = NotFound desc = could not find container \"35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594\": container with ID starting with 35231ee78b021fd72e84af197de71532352f303bb83514e6ef94ec4d98cb8594 not found: ID does not exist" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.191437 4907 scope.go:117] "RemoveContainer" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" Mar 13 16:20:30 crc kubenswrapper[4907]: E0313 16:20:30.191762 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85\": container with ID starting with b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85 not found: ID does not exist" containerID="b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85" Mar 13 16:20:30 crc kubenswrapper[4907]: I0313 16:20:30.191797 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85"} err="failed to get container status \"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85\": rpc error: code = NotFound desc = could not find container \"b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85\": container with ID starting with b5e33e84c8e3054fb9999d59d99976f5ddce20795949dbe8c0de722094ed3f85 not found: ID does not exist" Mar 13 16:20:31 crc kubenswrapper[4907]: I0313 16:20:31.791707 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:31 crc kubenswrapper[4907]: E0313 16:20:31.792114 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:31 crc kubenswrapper[4907]: I0313 16:20:31.798720 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" path="/var/lib/kubelet/pods/e879dfd4-a757-496a-9255-3905537d4c2f/volumes" Mar 13 16:20:42 crc kubenswrapper[4907]: I0313 16:20:42.181860 4907 scope.go:117] "RemoveContainer" containerID="048a0a595dbc6f4c8cb2226742f2beb5b639bec2395f93e0935301c145de8608" Mar 13 16:20:42 crc kubenswrapper[4907]: I0313 16:20:42.783315 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:42 crc kubenswrapper[4907]: E0313 16:20:42.784316 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:20:53 crc kubenswrapper[4907]: I0313 16:20:53.782648 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:20:54 crc kubenswrapper[4907]: I0313 16:20:54.325860 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31"} Mar 13 16:21:57 crc kubenswrapper[4907]: I0313 16:21:57.671722 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-564986d4bf-zhh4c" podUID="3ce663fe-b525-44ea-a6d4-33f6b2366f46" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.158063 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:22:00 crc kubenswrapper[4907]: E0313 16:22:00.159130 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159146 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" Mar 13 16:22:00 crc kubenswrapper[4907]: E0313 16:22:00.159180 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-utilities" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159190 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-utilities" Mar 13 16:22:00 crc kubenswrapper[4907]: E0313 16:22:00.159225 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-content" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159234 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="extract-content" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.159516 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e879dfd4-a757-496a-9255-3905537d4c2f" containerName="registry-server" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.160428 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.162595 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.162819 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.163865 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.167471 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.264348 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"auto-csr-approver-29556982-t6wrd\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.366114 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"auto-csr-approver-29556982-t6wrd\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.390584 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"auto-csr-approver-29556982-t6wrd\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.492787 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:00 crc kubenswrapper[4907]: I0313 16:22:00.970084 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:22:00 crc kubenswrapper[4907]: W0313 16:22:00.981851 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2364a186_f894_44a5_9c69_101cbfa14a82.slice/crio-86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a WatchSource:0}: Error finding container 86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a: Status 404 returned error can't find the container with id 86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a Mar 13 16:22:01 crc kubenswrapper[4907]: I0313 16:22:01.073674 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" event={"ID":"2364a186-f894-44a5-9c69-101cbfa14a82","Type":"ContainerStarted","Data":"86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a"} Mar 13 16:22:03 crc kubenswrapper[4907]: I0313 16:22:03.099076 4907 generic.go:334] "Generic (PLEG): container finished" podID="2364a186-f894-44a5-9c69-101cbfa14a82" containerID="48a114e32689f3100e68121daa48f2b4ba681c9f23fd2dfae93832f7ab1cf591" exitCode=0 Mar 13 16:22:03 crc kubenswrapper[4907]: I0313 16:22:03.099599 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" event={"ID":"2364a186-f894-44a5-9c69-101cbfa14a82","Type":"ContainerDied","Data":"48a114e32689f3100e68121daa48f2b4ba681c9f23fd2dfae93832f7ab1cf591"} Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.596297 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.669310 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") pod \"2364a186-f894-44a5-9c69-101cbfa14a82\" (UID: \"2364a186-f894-44a5-9c69-101cbfa14a82\") " Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.675784 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw" (OuterVolumeSpecName: "kube-api-access-mxsfw") pod "2364a186-f894-44a5-9c69-101cbfa14a82" (UID: "2364a186-f894-44a5-9c69-101cbfa14a82"). InnerVolumeSpecName "kube-api-access-mxsfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:22:04 crc kubenswrapper[4907]: I0313 16:22:04.772460 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxsfw\" (UniqueName: \"kubernetes.io/projected/2364a186-f894-44a5-9c69-101cbfa14a82-kube-api-access-mxsfw\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.124193 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" event={"ID":"2364a186-f894-44a5-9c69-101cbfa14a82","Type":"ContainerDied","Data":"86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a"} Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.124221 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556982-t6wrd" Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.124238 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86827b0005e293ad98b929277b369ca063a389d985875e7d08c7e28a85e0163a" Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.673237 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.680979 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556976-twcx4"] Mar 13 16:22:05 crc kubenswrapper[4907]: I0313 16:22:05.803775 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="362f626f-a4dd-45a7-ad88-4c539732eebb" path="/var/lib/kubelet/pods/362f626f-a4dd-45a7-ad88-4c539732eebb/volumes" Mar 13 16:22:42 crc kubenswrapper[4907]: I0313 16:22:42.329462 4907 scope.go:117] "RemoveContainer" containerID="7e0af7bb09c83c72fd3ad1550fb6bc5bec78d9ec9c71037ff20efe437fc967c9" Mar 13 16:22:50 crc kubenswrapper[4907]: I0313 16:22:50.607343 4907 generic.go:334] "Generic (PLEG): container finished" podID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerID="24ef5c64c1022adaf88de5eb0ff8963c4a40b5475aaa5f2ebc31c41febbdb0f6" exitCode=0 Mar 13 16:22:50 crc kubenswrapper[4907]: I0313 16:22:50.607421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerDied","Data":"24ef5c64c1022adaf88de5eb0ff8963c4a40b5475aaa5f2ebc31c41febbdb0f6"} Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.198814 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331072 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331132 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331217 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331322 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331344 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331365 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331382 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331420 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331470 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331502 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331554 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.331579 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") pod \"563c85ca-38c8-4853-a278-2a46fd0347a3\" (UID: \"563c85ca-38c8-4853-a278-2a46fd0347a3\") " Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.341645 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph" (OuterVolumeSpecName: "ceph") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.346615 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.347480 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k" (OuterVolumeSpecName: "kube-api-access-d896k") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "kube-api-access-d896k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.367673 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.367829 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.371926 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.372334 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.376262 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.385461 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.386462 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.406088 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory" (OuterVolumeSpecName: "inventory") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.407956 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.418747 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "563c85ca-38c8-4853-a278-2a46fd0347a3" (UID: "563c85ca-38c8-4853-a278-2a46fd0347a3"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.439808 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d896k\" (UniqueName: \"kubernetes.io/projected/563c85ca-38c8-4853-a278-2a46fd0347a3-kube-api-access-d896k\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.439853 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.439872 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440173 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440203 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440213 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440225 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440237 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440250 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440626 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440785 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440809 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.440822 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/563c85ca-38c8-4853-a278-2a46fd0347a3-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.635642 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" event={"ID":"563c85ca-38c8-4853-a278-2a46fd0347a3","Type":"ContainerDied","Data":"f3963807c8d98e582dff2a1a7b85c9b8817c6b50cae040491cc0e2d933ebfd1c"} Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.635699 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3963807c8d98e582dff2a1a7b85c9b8817c6b50cae040491cc0e2d933ebfd1c" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.635838 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-2swxs" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.813341 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-gmnh8"] Mar 13 16:22:52 crc kubenswrapper[4907]: E0313 16:22:52.813950 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerName="nova-cell1-openstack-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.813974 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerName="nova-cell1-openstack-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: E0313 16:22:52.814010 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" containerName="oc" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.814018 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" containerName="oc" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.814284 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="563c85ca-38c8-4853-a278-2a46fd0347a3" containerName="nova-cell1-openstack-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.814331 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" containerName="oc" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.815929 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.837130 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-gmnh8"] Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872381 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872439 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872613 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.872447 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.873743 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977712 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977800 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977823 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977864 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977907 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977929 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:52 crc kubenswrapper[4907]: I0313 16:22:52.977971 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083064 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083345 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083455 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083533 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083694 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083832 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.083920 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.096787 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.102475 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.104671 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.105261 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.105268 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.108554 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.114648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.124760 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"telemetry-openstack-openstack-cell1-gmnh8\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.209856 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.636251 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-gmnh8"] Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.639355 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:22:53 crc kubenswrapper[4907]: I0313 16:22:53.648673 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerStarted","Data":"7dfdeb24873be617cea682f0960f2b304c8f052d988caccff6de24f90a7a69e8"} Mar 13 16:22:54 crc kubenswrapper[4907]: I0313 16:22:54.659383 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerStarted","Data":"f1b9311ea4a0fb969131c4a9d82756d2d971316a7c3c03b8fb8bdcf2c37cd315"} Mar 13 16:22:54 crc kubenswrapper[4907]: I0313 16:22:54.686334 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" podStartSLOduration=2.107624687 podStartE2EDuration="2.686315194s" podCreationTimestamp="2026-03-13 16:22:52 +0000 UTC" firstStartedPulling="2026-03-13 16:22:53.639155655 +0000 UTC m=+8272.538943344" lastFinishedPulling="2026-03-13 16:22:54.217846142 +0000 UTC m=+8273.117633851" observedRunningTime="2026-03-13 16:22:54.681346177 +0000 UTC m=+8273.581133866" watchObservedRunningTime="2026-03-13 16:22:54.686315194 +0000 UTC m=+8273.586102883" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.185447 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.196660 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.256893 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.287058 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.287161 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.287199 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388422 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388524 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388555 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.388963 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.389100 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.436593 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"certified-operators-xgc75\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:05 crc kubenswrapper[4907]: I0313 16:23:05.546663 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.090535 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.798036 4907 generic.go:334] "Generic (PLEG): container finished" podID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" exitCode=0 Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.798186 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070"} Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.798498 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerStarted","Data":"9469a335dd59f1115dde7a5de517c050a6b96d5a4cf517b7778fd82b0e770bc9"} Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.987094 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:06 crc kubenswrapper[4907]: I0313 16:23:06.990668 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.015949 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.127539 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.127682 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.127764 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.229757 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.229965 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.230063 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.230466 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.230558 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.253044 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"community-operators-65t7j\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.314014 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:07 crc kubenswrapper[4907]: I0313 16:23:07.852556 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.821111 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerStarted","Data":"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468"} Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.823873 4907 generic.go:334] "Generic (PLEG): container finished" podID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" exitCode=0 Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.824039 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c"} Mar 13 16:23:08 crc kubenswrapper[4907]: I0313 16:23:08.824147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerStarted","Data":"8c53a2d922cabc54204bd5744de5a1de782425903b460f4444aa5c0d4c525729"} Mar 13 16:23:09 crc kubenswrapper[4907]: I0313 16:23:09.839057 4907 generic.go:334] "Generic (PLEG): container finished" podID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" exitCode=0 Mar 13 16:23:09 crc kubenswrapper[4907]: I0313 16:23:09.839172 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468"} Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.863225 4907 generic.go:334] "Generic (PLEG): container finished" podID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" exitCode=0 Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.863761 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55"} Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.867253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerStarted","Data":"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8"} Mar 13 16:23:10 crc kubenswrapper[4907]: I0313 16:23:10.923335 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xgc75" podStartSLOduration=2.303136071 podStartE2EDuration="5.923317688s" podCreationTimestamp="2026-03-13 16:23:05 +0000 UTC" firstStartedPulling="2026-03-13 16:23:06.80032306 +0000 UTC m=+8285.700110769" lastFinishedPulling="2026-03-13 16:23:10.420504667 +0000 UTC m=+8289.320292386" observedRunningTime="2026-03-13 16:23:10.910080996 +0000 UTC m=+8289.809868685" watchObservedRunningTime="2026-03-13 16:23:10.923317688 +0000 UTC m=+8289.823105377" Mar 13 16:23:11 crc kubenswrapper[4907]: I0313 16:23:11.880773 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerStarted","Data":"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0"} Mar 13 16:23:11 crc kubenswrapper[4907]: I0313 16:23:11.901459 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-65t7j" podStartSLOduration=3.233586485 podStartE2EDuration="5.901439458s" podCreationTimestamp="2026-03-13 16:23:06 +0000 UTC" firstStartedPulling="2026-03-13 16:23:08.82600973 +0000 UTC m=+8287.725797419" lastFinishedPulling="2026-03-13 16:23:11.493862693 +0000 UTC m=+8290.393650392" observedRunningTime="2026-03-13 16:23:11.899571427 +0000 UTC m=+8290.799359126" watchObservedRunningTime="2026-03-13 16:23:11.901439458 +0000 UTC m=+8290.801227157" Mar 13 16:23:15 crc kubenswrapper[4907]: I0313 16:23:15.547147 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:15 crc kubenswrapper[4907]: I0313 16:23:15.547751 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:15 crc kubenswrapper[4907]: I0313 16:23:15.639974 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:16 crc kubenswrapper[4907]: I0313 16:23:16.011586 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:17 crc kubenswrapper[4907]: I0313 16:23:17.314191 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:17 crc kubenswrapper[4907]: I0313 16:23:17.314244 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:17 crc kubenswrapper[4907]: I0313 16:23:17.373968 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:18 crc kubenswrapper[4907]: I0313 16:23:18.017549 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:18 crc kubenswrapper[4907]: I0313 16:23:18.041355 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:23:18 crc kubenswrapper[4907]: I0313 16:23:18.041424 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.403785 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.404309 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xgc75" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" containerID="cri-o://0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" gracePeriod=2 Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.921547 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.981918 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:20 crc kubenswrapper[4907]: I0313 16:23:20.982136 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-65t7j" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" containerID="cri-o://be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" gracePeriod=2 Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:20.999850 4907 generic.go:334] "Generic (PLEG): container finished" podID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" exitCode=0 Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:20.999972 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xgc75" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:20.999994 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8"} Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.000280 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xgc75" event={"ID":"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f","Type":"ContainerDied","Data":"9469a335dd59f1115dde7a5de517c050a6b96d5a4cf517b7778fd82b0e770bc9"} Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.000297 4907 scope.go:117] "RemoveContainer" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.020709 4907 scope.go:117] "RemoveContainer" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.036111 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") pod \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.036259 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") pod \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.036453 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") pod \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\" (UID: \"dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.042935 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities" (OuterVolumeSpecName: "utilities") pod "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" (UID: "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.050231 4907 scope.go:117] "RemoveContainer" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.050315 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5" (OuterVolumeSpecName: "kube-api-access-pqvp5") pod "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" (UID: "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f"). InnerVolumeSpecName "kube-api-access-pqvp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.089587 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" (UID: "dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.139104 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.139141 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqvp5\" (UniqueName: \"kubernetes.io/projected/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-kube-api-access-pqvp5\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.139154 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.210641 4907 scope.go:117] "RemoveContainer" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.211293 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8\": container with ID starting with 0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8 not found: ID does not exist" containerID="0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.211336 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8"} err="failed to get container status \"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8\": rpc error: code = NotFound desc = could not find container \"0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8\": container with ID starting with 0dd6c65f3084d0e1c74b304e1362948ff901486662743a1e17191174f6c000f8 not found: ID does not exist" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.211365 4907 scope.go:117] "RemoveContainer" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.212932 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468\": container with ID starting with 11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468 not found: ID does not exist" containerID="11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.212964 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468"} err="failed to get container status \"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468\": rpc error: code = NotFound desc = could not find container \"11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468\": container with ID starting with 11f2c82c12dc7835d9df366ec5b9750f88b9f2e5542cae61b51ff42bea344468 not found: ID does not exist" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.212983 4907 scope.go:117] "RemoveContainer" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.213248 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070\": container with ID starting with 28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070 not found: ID does not exist" containerID="28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.213278 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070"} err="failed to get container status \"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070\": rpc error: code = NotFound desc = could not find container \"28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070\": container with ID starting with 28de211c3a6fefc72bee431d7f78d30aea1198e921e5c1b38cddb8faf2098070 not found: ID does not exist" Mar 13 16:23:21 crc kubenswrapper[4907]: E0313 16:23:21.247324 4907 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07d0b87a_9167_4d87_8e03_ce6d6a804339.slice/crio-be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07d0b87a_9167_4d87_8e03_ce6d6a804339.slice/crio-conmon-be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0.scope\": RecentStats: unable to find data in memory cache]" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.340692 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.350222 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xgc75"] Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.414302 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.546938 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") pod \"07d0b87a-9167-4d87-8e03-ce6d6a804339\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.547262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") pod \"07d0b87a-9167-4d87-8e03-ce6d6a804339\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.547428 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") pod \"07d0b87a-9167-4d87-8e03-ce6d6a804339\" (UID: \"07d0b87a-9167-4d87-8e03-ce6d6a804339\") " Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.548230 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities" (OuterVolumeSpecName: "utilities") pod "07d0b87a-9167-4d87-8e03-ce6d6a804339" (UID: "07d0b87a-9167-4d87-8e03-ce6d6a804339"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.551289 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr" (OuterVolumeSpecName: "kube-api-access-xz4lr") pod "07d0b87a-9167-4d87-8e03-ce6d6a804339" (UID: "07d0b87a-9167-4d87-8e03-ce6d6a804339"). InnerVolumeSpecName "kube-api-access-xz4lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.599280 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07d0b87a-9167-4d87-8e03-ce6d6a804339" (UID: "07d0b87a-9167-4d87-8e03-ce6d6a804339"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.650042 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz4lr\" (UniqueName: \"kubernetes.io/projected/07d0b87a-9167-4d87-8e03-ce6d6a804339-kube-api-access-xz4lr\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.650075 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.650086 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07d0b87a-9167-4d87-8e03-ce6d6a804339-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:23:21 crc kubenswrapper[4907]: I0313 16:23:21.795026 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" path="/var/lib/kubelet/pods/dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f/volumes" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.017464 4907 generic.go:334] "Generic (PLEG): container finished" podID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" exitCode=0 Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.017517 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-65t7j" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.017535 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0"} Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.018043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-65t7j" event={"ID":"07d0b87a-9167-4d87-8e03-ce6d6a804339","Type":"ContainerDied","Data":"8c53a2d922cabc54204bd5744de5a1de782425903b460f4444aa5c0d4c525729"} Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.018070 4907 scope.go:117] "RemoveContainer" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.052218 4907 scope.go:117] "RemoveContainer" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.060294 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.073873 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-65t7j"] Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.084358 4907 scope.go:117] "RemoveContainer" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.145282 4907 scope.go:117] "RemoveContainer" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" Mar 13 16:23:22 crc kubenswrapper[4907]: E0313 16:23:22.145781 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0\": container with ID starting with be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0 not found: ID does not exist" containerID="be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.145832 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0"} err="failed to get container status \"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0\": rpc error: code = NotFound desc = could not find container \"be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0\": container with ID starting with be78711dbfea89545df6cec6b7c88e5079fe5c9b3bea035012fa58e156104ad0 not found: ID does not exist" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.145864 4907 scope.go:117] "RemoveContainer" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" Mar 13 16:23:22 crc kubenswrapper[4907]: E0313 16:23:22.146467 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55\": container with ID starting with ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55 not found: ID does not exist" containerID="ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.146497 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55"} err="failed to get container status \"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55\": rpc error: code = NotFound desc = could not find container \"ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55\": container with ID starting with ae44c65486353b068afdd09b8a8dcfc34f3a1a06614480a64352aea46afeaa55 not found: ID does not exist" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.146514 4907 scope.go:117] "RemoveContainer" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" Mar 13 16:23:22 crc kubenswrapper[4907]: E0313 16:23:22.147116 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c\": container with ID starting with ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c not found: ID does not exist" containerID="ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c" Mar 13 16:23:22 crc kubenswrapper[4907]: I0313 16:23:22.147177 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c"} err="failed to get container status \"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c\": rpc error: code = NotFound desc = could not find container \"ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c\": container with ID starting with ae19ff70cc9978ff9421f0487f5d5f5b18831382d5fa4d8e47971fb1fcf6259c not found: ID does not exist" Mar 13 16:23:23 crc kubenswrapper[4907]: I0313 16:23:23.812424 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" path="/var/lib/kubelet/pods/07d0b87a-9167-4d87-8e03-ce6d6a804339/volumes" Mar 13 16:23:48 crc kubenswrapper[4907]: I0313 16:23:48.041731 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:23:48 crc kubenswrapper[4907]: I0313 16:23:48.042439 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.163542 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164332 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164343 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164374 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164380 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164389 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164395 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164409 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164415 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="extract-utilities" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164429 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164434 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: E0313 16:24:00.164448 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164454 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="extract-content" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164629 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="07d0b87a-9167-4d87-8e03-ce6d6a804339" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.164653 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="dde3cb5c-6ddf-48c3-b6e1-5bb6a317076f" containerName="registry-server" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.165323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.167294 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.167407 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.171609 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.185296 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.258118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"auto-csr-approver-29556984-2lfjr\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.360249 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"auto-csr-approver-29556984-2lfjr\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.377742 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"auto-csr-approver-29556984-2lfjr\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.482971 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:00 crc kubenswrapper[4907]: I0313 16:24:00.972213 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:24:01 crc kubenswrapper[4907]: I0313 16:24:01.476091 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" event={"ID":"d0090a6a-15d8-4445-9830-9cab1aec7fdb","Type":"ContainerStarted","Data":"9efd579447758f2672ed2e2d8b1e7abfd2ad683604b38a7047ec29a50d072a9f"} Mar 13 16:24:02 crc kubenswrapper[4907]: I0313 16:24:02.489658 4907 generic.go:334] "Generic (PLEG): container finished" podID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerID="5fb59125e06d01b85c52baad33f05fd7f12c902ae1f36e7af1e62fdbcb0056b5" exitCode=0 Mar 13 16:24:02 crc kubenswrapper[4907]: I0313 16:24:02.489718 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" event={"ID":"d0090a6a-15d8-4445-9830-9cab1aec7fdb","Type":"ContainerDied","Data":"5fb59125e06d01b85c52baad33f05fd7f12c902ae1f36e7af1e62fdbcb0056b5"} Mar 13 16:24:03 crc kubenswrapper[4907]: I0313 16:24:03.916290 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:03 crc kubenswrapper[4907]: I0313 16:24:03.955989 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") pod \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\" (UID: \"d0090a6a-15d8-4445-9830-9cab1aec7fdb\") " Mar 13 16:24:03 crc kubenswrapper[4907]: I0313 16:24:03.964904 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6" (OuterVolumeSpecName: "kube-api-access-655b6") pod "d0090a6a-15d8-4445-9830-9cab1aec7fdb" (UID: "d0090a6a-15d8-4445-9830-9cab1aec7fdb"). InnerVolumeSpecName "kube-api-access-655b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.058856 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-655b6\" (UniqueName: \"kubernetes.io/projected/d0090a6a-15d8-4445-9830-9cab1aec7fdb-kube-api-access-655b6\") on node \"crc\" DevicePath \"\"" Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.517411 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" event={"ID":"d0090a6a-15d8-4445-9830-9cab1aec7fdb","Type":"ContainerDied","Data":"9efd579447758f2672ed2e2d8b1e7abfd2ad683604b38a7047ec29a50d072a9f"} Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.517454 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9efd579447758f2672ed2e2d8b1e7abfd2ad683604b38a7047ec29a50d072a9f" Mar 13 16:24:04 crc kubenswrapper[4907]: I0313 16:24:04.517475 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556984-2lfjr" Mar 13 16:24:05 crc kubenswrapper[4907]: I0313 16:24:05.005594 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:24:05 crc kubenswrapper[4907]: I0313 16:24:05.014848 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556978-m5zzz"] Mar 13 16:24:05 crc kubenswrapper[4907]: I0313 16:24:05.829446 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1" path="/var/lib/kubelet/pods/b4a302e0-25a2-43d0-b5c0-5047d7a2c5a1/volumes" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.042407 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.043205 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.043268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.044518 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.044601 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31" gracePeriod=600 Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.673920 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31" exitCode=0 Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.674140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31"} Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.674376 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69"} Mar 13 16:24:18 crc kubenswrapper[4907]: I0313 16:24:18.674411 4907 scope.go:117] "RemoveContainer" containerID="300e429cce24fa4eb9aefff9494fe9905ee7b0767504365170f1f9645a3c8b5f" Mar 13 16:24:42 crc kubenswrapper[4907]: I0313 16:24:42.493791 4907 scope.go:117] "RemoveContainer" containerID="f49981cfe3caeb1a9f80b1991ed24d10adb510db888dce88031b1e2cc8045b1b" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.168240 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:26:00 crc kubenswrapper[4907]: E0313 16:26:00.169546 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerName="oc" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.169569 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerName="oc" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.169930 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" containerName="oc" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.171293 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.174226 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.179777 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.180045 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.212093 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.213695 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"auto-csr-approver-29556986-nlfck\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.315765 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"auto-csr-approver-29556986-nlfck\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.342531 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"auto-csr-approver-29556986-nlfck\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:00 crc kubenswrapper[4907]: I0313 16:26:00.497323 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:01 crc kubenswrapper[4907]: I0313 16:26:01.025715 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:26:01 crc kubenswrapper[4907]: I0313 16:26:01.974847 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556986-nlfck" event={"ID":"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4","Type":"ContainerStarted","Data":"802d4a1e6437ddc0b99366dc5755b1fe26de2d7d7692ce9bb52745e8afad2ea6"} Mar 13 16:26:02 crc kubenswrapper[4907]: I0313 16:26:02.990051 4907 generic.go:334] "Generic (PLEG): container finished" podID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerID="4fc188ddf095682c93e7e1f3f3249b322a5fe91043eda372761cb774c37d43db" exitCode=0 Mar 13 16:26:02 crc kubenswrapper[4907]: I0313 16:26:02.990136 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556986-nlfck" event={"ID":"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4","Type":"ContainerDied","Data":"4fc188ddf095682c93e7e1f3f3249b322a5fe91043eda372761cb774c37d43db"} Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.489997 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.530005 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") pod \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\" (UID: \"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4\") " Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.547196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm" (OuterVolumeSpecName: "kube-api-access-zbssm") pod "1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" (UID: "1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4"). InnerVolumeSpecName "kube-api-access-zbssm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:26:04 crc kubenswrapper[4907]: I0313 16:26:04.631787 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbssm\" (UniqueName: \"kubernetes.io/projected/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4-kube-api-access-zbssm\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.023995 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556986-nlfck" event={"ID":"1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4","Type":"ContainerDied","Data":"802d4a1e6437ddc0b99366dc5755b1fe26de2d7d7692ce9bb52745e8afad2ea6"} Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.024044 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="802d4a1e6437ddc0b99366dc5755b1fe26de2d7d7692ce9bb52745e8afad2ea6" Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.024113 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556986-nlfck" Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.574973 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.586184 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556980-5b2q4"] Mar 13 16:26:05 crc kubenswrapper[4907]: I0313 16:26:05.797251 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b9615a4-0270-4660-bd34-5efded647a80" path="/var/lib/kubelet/pods/0b9615a4-0270-4660-bd34-5efded647a80/volumes" Mar 13 16:26:18 crc kubenswrapper[4907]: I0313 16:26:18.041722 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:26:18 crc kubenswrapper[4907]: I0313 16:26:18.042404 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:26:27 crc kubenswrapper[4907]: I0313 16:26:27.322134 4907 generic.go:334] "Generic (PLEG): container finished" podID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerID="f1b9311ea4a0fb969131c4a9d82756d2d971316a7c3c03b8fb8bdcf2c37cd315" exitCode=0 Mar 13 16:26:27 crc kubenswrapper[4907]: I0313 16:26:27.322223 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerDied","Data":"f1b9311ea4a0fb969131c4a9d82756d2d971316a7c3c03b8fb8bdcf2c37cd315"} Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.851693 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.869742 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.872577 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.872753 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.872863 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.873029 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.873242 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.873837 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.874260 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") pod \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\" (UID: \"b5718b26-a6b1-4b64-812b-a85da3d6f0a4\") " Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.886859 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph" (OuterVolumeSpecName: "ceph") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.884936 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.918499 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.929031 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z" (OuterVolumeSpecName: "kube-api-access-zkd9z") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "kube-api-access-zkd9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.932805 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.934296 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.935056 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.949644 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory" (OuterVolumeSpecName: "inventory") pod "b5718b26-a6b1-4b64-812b-a85da3d6f0a4" (UID: "b5718b26-a6b1-4b64-812b-a85da3d6f0a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.978982 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979223 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979359 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979471 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979580 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979688 4907 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.979949 4907 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:28 crc kubenswrapper[4907]: I0313 16:26:28.980251 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkd9z\" (UniqueName: \"kubernetes.io/projected/b5718b26-a6b1-4b64-812b-a85da3d6f0a4-kube-api-access-zkd9z\") on node \"crc\" DevicePath \"\"" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.353371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" event={"ID":"b5718b26-a6b1-4b64-812b-a85da3d6f0a4","Type":"ContainerDied","Data":"7dfdeb24873be617cea682f0960f2b304c8f052d988caccff6de24f90a7a69e8"} Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.353414 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dfdeb24873be617cea682f0960f2b304c8f052d988caccff6de24f90a7a69e8" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.353477 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-gmnh8" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.459415 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-sqxhp"] Mar 13 16:26:29 crc kubenswrapper[4907]: E0313 16:26:29.460005 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerName="telemetry-openstack-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460034 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerName="telemetry-openstack-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: E0313 16:26:29.460069 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerName="oc" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460078 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerName="oc" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460327 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" containerName="oc" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.460772 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5718b26-a6b1-4b64-812b-a85da3d6f0a4" containerName="telemetry-openstack-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.464934 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.466796 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.467364 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.467382 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.469202 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.469847 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490049 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490118 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490141 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490167 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490392 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.490529 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.497747 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-sqxhp"] Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592584 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592674 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592791 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592824 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592846 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.592874 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.596432 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.597018 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.597797 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.598058 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.598512 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.612260 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"neutron-sriov-openstack-openstack-cell1-sqxhp\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:29 crc kubenswrapper[4907]: I0313 16:26:29.783468 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:26:30 crc kubenswrapper[4907]: I0313 16:26:30.316542 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-sqxhp"] Mar 13 16:26:30 crc kubenswrapper[4907]: I0313 16:26:30.369760 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerStarted","Data":"762fb1ddd2d8fd925d29b186df5f88568978b2b1ec0e68aa4f06e19c4e53f86e"} Mar 13 16:26:31 crc kubenswrapper[4907]: I0313 16:26:31.387602 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerStarted","Data":"259cfb9d82d9b9b7b45bab20327a4d8c2e39c52001f26a91569b8c3052ad5700"} Mar 13 16:26:31 crc kubenswrapper[4907]: I0313 16:26:31.425771 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" podStartSLOduration=1.8257704860000001 podStartE2EDuration="2.425743676s" podCreationTimestamp="2026-03-13 16:26:29 +0000 UTC" firstStartedPulling="2026-03-13 16:26:30.325658588 +0000 UTC m=+8489.225446277" lastFinishedPulling="2026-03-13 16:26:30.925631748 +0000 UTC m=+8489.825419467" observedRunningTime="2026-03-13 16:26:31.406685244 +0000 UTC m=+8490.306472973" watchObservedRunningTime="2026-03-13 16:26:31.425743676 +0000 UTC m=+8490.325531405" Mar 13 16:26:42 crc kubenswrapper[4907]: I0313 16:26:42.661215 4907 scope.go:117] "RemoveContainer" containerID="d626b98830070c30af12d3ea7ba374e64df5c680e5d7d98055b77c7c0b8bf280" Mar 13 16:26:48 crc kubenswrapper[4907]: I0313 16:26:48.041985 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:26:48 crc kubenswrapper[4907]: I0313 16:26:48.042934 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.043018 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.043768 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.043839 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.045244 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.045349 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" gracePeriod=600 Mar 13 16:27:18 crc kubenswrapper[4907]: E0313 16:27:18.169454 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.948802 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" exitCode=0 Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.948898 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69"} Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.949215 4907 scope.go:117] "RemoveContainer" containerID="50d5a864941e0a2162babd980c37981cbcd4e9bfc7e1598617888cc1ec5d9b31" Mar 13 16:27:18 crc kubenswrapper[4907]: I0313 16:27:18.950018 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:18 crc kubenswrapper[4907]: E0313 16:27:18.950444 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:33 crc kubenswrapper[4907]: I0313 16:27:33.783759 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:33 crc kubenswrapper[4907]: E0313 16:27:33.785055 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:40 crc kubenswrapper[4907]: I0313 16:27:40.198097 4907 generic.go:334] "Generic (PLEG): container finished" podID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerID="259cfb9d82d9b9b7b45bab20327a4d8c2e39c52001f26a91569b8c3052ad5700" exitCode=0 Mar 13 16:27:40 crc kubenswrapper[4907]: I0313 16:27:40.198239 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerDied","Data":"259cfb9d82d9b9b7b45bab20327a4d8c2e39c52001f26a91569b8c3052ad5700"} Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.724518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736543 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736701 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736786 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.736856 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.737063 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.737169 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") pod \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\" (UID: \"05d384d7-9635-4d68-8eaf-48fea8bed0b8\") " Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.745986 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.747047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6" (OuterVolumeSpecName: "kube-api-access-jl9p6") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "kube-api-access-jl9p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.762449 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph" (OuterVolumeSpecName: "ceph") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.793013 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.796348 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory" (OuterVolumeSpecName: "inventory") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.814815 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "05d384d7-9635-4d68-8eaf-48fea8bed0b8" (UID: "05d384d7-9635-4d68-8eaf-48fea8bed0b8"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849474 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849538 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl9p6\" (UniqueName: \"kubernetes.io/projected/05d384d7-9635-4d68-8eaf-48fea8bed0b8-kube-api-access-jl9p6\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849555 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849567 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849579 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:41 crc kubenswrapper[4907]: I0313 16:27:41.849614 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/05d384d7-9635-4d68-8eaf-48fea8bed0b8-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.228140 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" event={"ID":"05d384d7-9635-4d68-8eaf-48fea8bed0b8","Type":"ContainerDied","Data":"762fb1ddd2d8fd925d29b186df5f88568978b2b1ec0e68aa4f06e19c4e53f86e"} Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.228176 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-sqxhp" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.228182 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="762fb1ddd2d8fd925d29b186df5f88568978b2b1ec0e68aa4f06e19c4e53f86e" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.373174 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jmbms"] Mar 13 16:27:42 crc kubenswrapper[4907]: E0313 16:27:42.376598 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerName="neutron-sriov-openstack-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.376755 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerName="neutron-sriov-openstack-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.377149 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="05d384d7-9635-4d68-8eaf-48fea8bed0b8" containerName="neutron-sriov-openstack-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.378164 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.381954 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.382240 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.385435 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.385440 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.385976 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.395705 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jmbms"] Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.464422 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.464640 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.464846 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.465004 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.465146 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.465207 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.566878 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.566997 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567042 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567124 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.567150 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.572074 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.572622 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.573628 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.579267 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.580008 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.600514 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"neutron-dhcp-openstack-openstack-cell1-jmbms\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:42 crc kubenswrapper[4907]: I0313 16:27:42.710570 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:27:43 crc kubenswrapper[4907]: I0313 16:27:43.362405 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-jmbms"] Mar 13 16:27:43 crc kubenswrapper[4907]: W0313 16:27:43.365916 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5764492_ad4a_4e43_8a98_99ea7cfd9248.slice/crio-ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32 WatchSource:0}: Error finding container ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32: Status 404 returned error can't find the container with id ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32 Mar 13 16:27:44 crc kubenswrapper[4907]: I0313 16:27:44.256238 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerStarted","Data":"0e4b9902269aa35ede42d672cbbcad84fd72d75aa3d82966fa7bf7b01617b050"} Mar 13 16:27:44 crc kubenswrapper[4907]: I0313 16:27:44.256923 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerStarted","Data":"ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32"} Mar 13 16:27:44 crc kubenswrapper[4907]: I0313 16:27:44.284023 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" podStartSLOduration=1.6693486210000001 podStartE2EDuration="2.284006223s" podCreationTimestamp="2026-03-13 16:27:42 +0000 UTC" firstStartedPulling="2026-03-13 16:27:43.369038892 +0000 UTC m=+8562.268826581" lastFinishedPulling="2026-03-13 16:27:43.983696464 +0000 UTC m=+8562.883484183" observedRunningTime="2026-03-13 16:27:44.279127119 +0000 UTC m=+8563.178914828" watchObservedRunningTime="2026-03-13 16:27:44.284006223 +0000 UTC m=+8563.183793912" Mar 13 16:27:45 crc kubenswrapper[4907]: I0313 16:27:45.782621 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:45 crc kubenswrapper[4907]: E0313 16:27:45.783291 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:27:59 crc kubenswrapper[4907]: I0313 16:27:59.782847 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:27:59 crc kubenswrapper[4907]: E0313 16:27:59.783870 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.156619 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.158229 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.160404 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.161774 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.162012 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.183126 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.292921 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"auto-csr-approver-29556988-t5tgt\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.394628 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"auto-csr-approver-29556988-t5tgt\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.425552 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"auto-csr-approver-29556988-t5tgt\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.485534 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.971090 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:28:00 crc kubenswrapper[4907]: I0313 16:28:00.985565 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:28:01 crc kubenswrapper[4907]: I0313 16:28:01.478195 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerStarted","Data":"f9ca69cff36c9453cd1eb3650007fe40309486ddb121cbafcc1bfcc8520c31ba"} Mar 13 16:28:02 crc kubenswrapper[4907]: I0313 16:28:02.490401 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerStarted","Data":"68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c"} Mar 13 16:28:02 crc kubenswrapper[4907]: I0313 16:28:02.514834 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" podStartSLOduration=1.494269263 podStartE2EDuration="2.514815854s" podCreationTimestamp="2026-03-13 16:28:00 +0000 UTC" firstStartedPulling="2026-03-13 16:28:00.984854603 +0000 UTC m=+8579.884642332" lastFinishedPulling="2026-03-13 16:28:02.005401234 +0000 UTC m=+8580.905188923" observedRunningTime="2026-03-13 16:28:02.505434708 +0000 UTC m=+8581.405222397" watchObservedRunningTime="2026-03-13 16:28:02.514815854 +0000 UTC m=+8581.414603543" Mar 13 16:28:03 crc kubenswrapper[4907]: I0313 16:28:03.506479 4907 generic.go:334] "Generic (PLEG): container finished" podID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerID="68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c" exitCode=0 Mar 13 16:28:03 crc kubenswrapper[4907]: I0313 16:28:03.506573 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerDied","Data":"68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c"} Mar 13 16:28:04 crc kubenswrapper[4907]: I0313 16:28:04.956299 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.028865 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") pod \"6aa9f963-9a2b-4101-88e5-9553783a9963\" (UID: \"6aa9f963-9a2b-4101-88e5-9553783a9963\") " Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.051178 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd" (OuterVolumeSpecName: "kube-api-access-2p6nd") pod "6aa9f963-9a2b-4101-88e5-9553783a9963" (UID: "6aa9f963-9a2b-4101-88e5-9553783a9963"). InnerVolumeSpecName "kube-api-access-2p6nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.131116 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p6nd\" (UniqueName: \"kubernetes.io/projected/6aa9f963-9a2b-4101-88e5-9553783a9963-kube-api-access-2p6nd\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.530816 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" event={"ID":"6aa9f963-9a2b-4101-88e5-9553783a9963","Type":"ContainerDied","Data":"f9ca69cff36c9453cd1eb3650007fe40309486ddb121cbafcc1bfcc8520c31ba"} Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.530852 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9ca69cff36c9453cd1eb3650007fe40309486ddb121cbafcc1bfcc8520c31ba" Mar 13 16:28:05 crc kubenswrapper[4907]: I0313 16:28:05.530938 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556988-t5tgt" Mar 13 16:28:06 crc kubenswrapper[4907]: I0313 16:28:06.048426 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:28:06 crc kubenswrapper[4907]: I0313 16:28:06.062301 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556982-t6wrd"] Mar 13 16:28:07 crc kubenswrapper[4907]: I0313 16:28:07.806026 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2364a186-f894-44a5-9c69-101cbfa14a82" path="/var/lib/kubelet/pods/2364a186-f894-44a5-9c69-101cbfa14a82/volumes" Mar 13 16:28:13 crc kubenswrapper[4907]: I0313 16:28:13.783290 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:13 crc kubenswrapper[4907]: E0313 16:28:13.784541 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:26 crc kubenswrapper[4907]: I0313 16:28:26.782381 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:26 crc kubenswrapper[4907]: E0313 16:28:26.783289 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.818157 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:33 crc kubenswrapper[4907]: E0313 16:28:33.819026 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerName="oc" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.819038 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerName="oc" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.819235 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" containerName="oc" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.820700 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.839365 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.958948 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.958992 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:33 crc kubenswrapper[4907]: I0313 16:28:33.959376 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.061445 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.061537 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.061752 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.062023 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.062210 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.089147 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"redhat-marketplace-96kxt\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.139440 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.692325 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:34 crc kubenswrapper[4907]: I0313 16:28:34.898194 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerStarted","Data":"0dafac436ece1b2ddb45a68c7f74b1a3663a57e9df7fd821ecccf3252eb8e180"} Mar 13 16:28:35 crc kubenswrapper[4907]: I0313 16:28:35.907522 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" exitCode=0 Mar 13 16:28:35 crc kubenswrapper[4907]: I0313 16:28:35.907734 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170"} Mar 13 16:28:37 crc kubenswrapper[4907]: I0313 16:28:37.936019 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerStarted","Data":"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864"} Mar 13 16:28:38 crc kubenswrapper[4907]: I0313 16:28:38.949452 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" exitCode=0 Mar 13 16:28:38 crc kubenswrapper[4907]: I0313 16:28:38.949551 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864"} Mar 13 16:28:40 crc kubenswrapper[4907]: I0313 16:28:40.783311 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:40 crc kubenswrapper[4907]: E0313 16:28:40.784109 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:28:41 crc kubenswrapper[4907]: I0313 16:28:41.011778 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerStarted","Data":"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6"} Mar 13 16:28:41 crc kubenswrapper[4907]: I0313 16:28:41.047570 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-96kxt" podStartSLOduration=4.519417311 podStartE2EDuration="8.047537196s" podCreationTimestamp="2026-03-13 16:28:33 +0000 UTC" firstStartedPulling="2026-03-13 16:28:35.909557751 +0000 UTC m=+8614.809345440" lastFinishedPulling="2026-03-13 16:28:39.437677626 +0000 UTC m=+8618.337465325" observedRunningTime="2026-03-13 16:28:41.039721745 +0000 UTC m=+8619.939509474" watchObservedRunningTime="2026-03-13 16:28:41.047537196 +0000 UTC m=+8619.947324925" Mar 13 16:28:42 crc kubenswrapper[4907]: I0313 16:28:42.781020 4907 scope.go:117] "RemoveContainer" containerID="48a114e32689f3100e68121daa48f2b4ba681c9f23fd2dfae93832f7ab1cf591" Mar 13 16:28:44 crc kubenswrapper[4907]: I0313 16:28:44.140115 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:44 crc kubenswrapper[4907]: I0313 16:28:44.140545 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:44 crc kubenswrapper[4907]: I0313 16:28:44.204611 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:45 crc kubenswrapper[4907]: I0313 16:28:45.117603 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:45 crc kubenswrapper[4907]: I0313 16:28:45.176589 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.081225 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-96kxt" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" containerID="cri-o://15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" gracePeriod=2 Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.603759 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.623621 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") pod \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.623687 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") pod \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.623733 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") pod \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\" (UID: \"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748\") " Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.624738 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities" (OuterVolumeSpecName: "utilities") pod "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" (UID: "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.633620 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp" (OuterVolumeSpecName: "kube-api-access-zg9rp") pod "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" (UID: "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748"). InnerVolumeSpecName "kube-api-access-zg9rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.668248 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" (UID: "1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.726291 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg9rp\" (UniqueName: \"kubernetes.io/projected/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-kube-api-access-zg9rp\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.726345 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:47 crc kubenswrapper[4907]: I0313 16:28:47.726366 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127088 4907 generic.go:334] "Generic (PLEG): container finished" podID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" exitCode=0 Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127459 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-96kxt" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127544 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6"} Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127579 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-96kxt" event={"ID":"1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748","Type":"ContainerDied","Data":"0dafac436ece1b2ddb45a68c7f74b1a3663a57e9df7fd821ecccf3252eb8e180"} Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.127601 4907 scope.go:117] "RemoveContainer" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.164456 4907 scope.go:117] "RemoveContainer" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.170028 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.178851 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-96kxt"] Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.197461 4907 scope.go:117] "RemoveContainer" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.239390 4907 scope.go:117] "RemoveContainer" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" Mar 13 16:28:48 crc kubenswrapper[4907]: E0313 16:28:48.239904 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6\": container with ID starting with 15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6 not found: ID does not exist" containerID="15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.239989 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6"} err="failed to get container status \"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6\": rpc error: code = NotFound desc = could not find container \"15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6\": container with ID starting with 15dbe6019db9afdf340463b89d9f2149f147e6017e7c5bec357846162e1071e6 not found: ID does not exist" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.240021 4907 scope.go:117] "RemoveContainer" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" Mar 13 16:28:48 crc kubenswrapper[4907]: E0313 16:28:48.240543 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864\": container with ID starting with 9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864 not found: ID does not exist" containerID="9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.240635 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864"} err="failed to get container status \"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864\": rpc error: code = NotFound desc = could not find container \"9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864\": container with ID starting with 9c0f31afbd3f516e339a816053788dfb8af39539535629431e6e575f2db70864 not found: ID does not exist" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.240669 4907 scope.go:117] "RemoveContainer" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" Mar 13 16:28:48 crc kubenswrapper[4907]: E0313 16:28:48.241204 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170\": container with ID starting with fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170 not found: ID does not exist" containerID="fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170" Mar 13 16:28:48 crc kubenswrapper[4907]: I0313 16:28:48.241251 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170"} err="failed to get container status \"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170\": rpc error: code = NotFound desc = could not find container \"fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170\": container with ID starting with fc7c8ee006b2576371976490d6643daa901bc4a5c44c8411005168ae914f4170 not found: ID does not exist" Mar 13 16:28:49 crc kubenswrapper[4907]: I0313 16:28:49.794153 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" path="/var/lib/kubelet/pods/1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748/volumes" Mar 13 16:28:51 crc kubenswrapper[4907]: I0313 16:28:51.796631 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:28:51 crc kubenswrapper[4907]: E0313 16:28:51.797665 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:01 crc kubenswrapper[4907]: I0313 16:29:01.303450 4907 generic.go:334] "Generic (PLEG): container finished" podID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerID="0e4b9902269aa35ede42d672cbbcad84fd72d75aa3d82966fa7bf7b01617b050" exitCode=0 Mar 13 16:29:01 crc kubenswrapper[4907]: I0313 16:29:01.303657 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerDied","Data":"0e4b9902269aa35ede42d672cbbcad84fd72d75aa3d82966fa7bf7b01617b050"} Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.783854 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.887724 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888133 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888222 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888315 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888358 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.888394 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") pod \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\" (UID: \"b5764492-ad4a-4e43-8a98-99ea7cfd9248\") " Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.893758 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz" (OuterVolumeSpecName: "kube-api-access-dw6mz") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "kube-api-access-dw6mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.894047 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.899748 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph" (OuterVolumeSpecName: "ceph") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.916717 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory" (OuterVolumeSpecName: "inventory") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.929066 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.940341 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "b5764492-ad4a-4e43-8a98-99ea7cfd9248" (UID: "b5764492-ad4a-4e43-8a98-99ea7cfd9248"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.990779 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.990935 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.990999 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw6mz\" (UniqueName: \"kubernetes.io/projected/b5764492-ad4a-4e43-8a98-99ea7cfd9248-kube-api-access-dw6mz\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.991068 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.991129 4907 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:02 crc kubenswrapper[4907]: I0313 16:29:02.991186 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/b5764492-ad4a-4e43-8a98-99ea7cfd9248-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:03 crc kubenswrapper[4907]: I0313 16:29:03.335975 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" event={"ID":"b5764492-ad4a-4e43-8a98-99ea7cfd9248","Type":"ContainerDied","Data":"ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32"} Mar 13 16:29:03 crc kubenswrapper[4907]: I0313 16:29:03.336333 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba830f46ef1e05ea0e668ef0942e0f24774e231102347a25077be9ae94ffda32" Mar 13 16:29:03 crc kubenswrapper[4907]: I0313 16:29:03.336092 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-jmbms" Mar 13 16:29:06 crc kubenswrapper[4907]: I0313 16:29:06.783278 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:06 crc kubenswrapper[4907]: E0313 16:29:06.784186 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:21 crc kubenswrapper[4907]: I0313 16:29:21.795572 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:21 crc kubenswrapper[4907]: E0313 16:29:21.796791 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:26 crc kubenswrapper[4907]: I0313 16:29:26.506181 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:26 crc kubenswrapper[4907]: I0313 16:29:26.506785 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" containerID="cri-o://c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.041898 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.042111 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.195818 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.196343 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" containerID="cri-o://5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.216861 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.217145 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" containerID="cri-o://fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.217219 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" containerID="cri-o://19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.231168 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.231502 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" containerID="cri-o://047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.232184 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" containerID="cri-o://c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d" gracePeriod=30 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.309656 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc"] Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310050 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerName="neutron-dhcp-openstack-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310068 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerName="neutron-dhcp-openstack-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310091 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310098 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310120 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-utilities" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310126 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-utilities" Mar 13 16:29:27 crc kubenswrapper[4907]: E0313 16:29:27.310148 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-content" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310154 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="extract-content" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310341 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5764492-ad4a-4e43-8a98-99ea7cfd9248" containerName="neutron-dhcp-openstack-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.310364 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ba75cd1-2e4a-4695-b33c-8ed0cbbbb748" containerName="registry-server" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.311074 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.316001 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317117 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317308 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317788 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317867 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.317911 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-44kxw" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.320113 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.335702 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc"] Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385395 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385457 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385484 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385643 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385690 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385861 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.385985 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.386725 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.386930 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.386996 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.387061 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.387098 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489248 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489500 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489607 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489707 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489802 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.489922 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490008 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490091 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490198 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490288 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490387 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490480 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.490562 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.493224 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.495758 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.502530 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.505648 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.509424 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.509603 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.509705 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.510086 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.511025 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.514447 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.514465 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.515674 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.534664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.624521 4907 generic.go:334] "Generic (PLEG): container finished" podID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerID="c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d" exitCode=143 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.624609 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerDied","Data":"c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d"} Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.627020 4907 generic.go:334] "Generic (PLEG): container finished" podID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerID="fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861" exitCode=143 Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.627099 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerDied","Data":"fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861"} Mar 13 16:29:27 crc kubenswrapper[4907]: I0313 16:29:27.730153 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.305328 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.307162 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.309043 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 16:29:28 crc kubenswrapper[4907]: E0313 16:29:28.309095 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:28 crc kubenswrapper[4907]: I0313 16:29:28.331603 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc"] Mar 13 16:29:28 crc kubenswrapper[4907]: I0313 16:29:28.638387 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerStarted","Data":"315fdd9184d5c2a2780c5ff803490daf8312afa1aa0a6d15f69dbc1f771bf80c"} Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.265093 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.271375 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.273580 4907 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 16:29:29 crc kubenswrapper[4907]: E0313 16:29:29.273634 4907 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.653007 4907 generic.go:334] "Generic (PLEG): container finished" podID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerID="c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe" exitCode=0 Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.653110 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerDied","Data":"c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe"} Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.655297 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerStarted","Data":"8644ffba95037a0d279e1ad31f88e2f9c3fae28fb5fd51859c1fb3e25cec04cc"} Mar 13 16:29:29 crc kubenswrapper[4907]: I0313 16:29:29.694788 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" podStartSLOduration=1.983359394 podStartE2EDuration="2.694773164s" podCreationTimestamp="2026-03-13 16:29:27 +0000 UTC" firstStartedPulling="2026-03-13 16:29:28.337268342 +0000 UTC m=+8667.237056031" lastFinishedPulling="2026-03-13 16:29:29.048682082 +0000 UTC m=+8667.948469801" observedRunningTime="2026-03-13 16:29:29.680929935 +0000 UTC m=+8668.580717624" watchObservedRunningTime="2026-03-13 16:29:29.694773164 +0000 UTC m=+8668.594560853" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.037734 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.156667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") pod \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.156955 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") pod \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.157007 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") pod \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\" (UID: \"647720e3-d21e-4f41-aeb8-d4e8f092cbb5\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.162951 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv" (OuterVolumeSpecName: "kube-api-access-s7nfv") pod "647720e3-d21e-4f41-aeb8-d4e8f092cbb5" (UID: "647720e3-d21e-4f41-aeb8-d4e8f092cbb5"). InnerVolumeSpecName "kube-api-access-s7nfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.185026 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data" (OuterVolumeSpecName: "config-data") pod "647720e3-d21e-4f41-aeb8-d4e8f092cbb5" (UID: "647720e3-d21e-4f41-aeb8-d4e8f092cbb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.196546 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "647720e3-d21e-4f41-aeb8-d4e8f092cbb5" (UID: "647720e3-d21e-4f41-aeb8-d4e8f092cbb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.208270 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.259573 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.259604 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.259615 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7nfv\" (UniqueName: \"kubernetes.io/projected/647720e3-d21e-4f41-aeb8-d4e8f092cbb5-kube-api-access-s7nfv\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.360899 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") pod \"eded65eb-7a04-4f90-ad82-be65db9014eb\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.361671 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") pod \"eded65eb-7a04-4f90-ad82-be65db9014eb\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.361850 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") pod \"eded65eb-7a04-4f90-ad82-be65db9014eb\" (UID: \"eded65eb-7a04-4f90-ad82-be65db9014eb\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.365100 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc" (OuterVolumeSpecName: "kube-api-access-hh9bc") pod "eded65eb-7a04-4f90-ad82-be65db9014eb" (UID: "eded65eb-7a04-4f90-ad82-be65db9014eb"). InnerVolumeSpecName "kube-api-access-hh9bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.393972 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data" (OuterVolumeSpecName: "config-data") pod "eded65eb-7a04-4f90-ad82-be65db9014eb" (UID: "eded65eb-7a04-4f90-ad82-be65db9014eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.397010 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eded65eb-7a04-4f90-ad82-be65db9014eb" (UID: "eded65eb-7a04-4f90-ad82-be65db9014eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.464189 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.464224 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh9bc\" (UniqueName: \"kubernetes.io/projected/eded65eb-7a04-4f90-ad82-be65db9014eb-kube-api-access-hh9bc\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.464249 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eded65eb-7a04-4f90-ad82-be65db9014eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.667155 4907 generic.go:334] "Generic (PLEG): container finished" podID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerID="19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a" exitCode=0 Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.667476 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerDied","Data":"19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.669369 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.669388 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"647720e3-d21e-4f41-aeb8-d4e8f092cbb5","Type":"ContainerDied","Data":"b4d79e83e6b122d073a9e884515b0623866f06609025b076fcf8513ad8311abc"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.669441 4907 scope.go:117] "RemoveContainer" containerID="c20788b6b1aecba8b32acff82d1977637a4b4a170211b700ca1e5ce58e45cffe" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672029 4907 generic.go:334] "Generic (PLEG): container finished" podID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" exitCode=0 Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672082 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerDied","Data":"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672102 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"eded65eb-7a04-4f90-ad82-be65db9014eb","Type":"ContainerDied","Data":"4c8aa8b2ae3cb083bba62c66c737dee676b8b626c0c8f53960f20765d91bb7b2"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.672156 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.676986 4907 generic.go:334] "Generic (PLEG): container finished" podID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerID="047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4" exitCode=0 Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.677043 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerDied","Data":"047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4"} Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.741893 4907 scope.go:117] "RemoveContainer" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.772281 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.823864 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.834093 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878077 4907 scope.go:117] "RemoveContainer" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.878450 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d\": container with ID starting with e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d not found: ID does not exist" containerID="e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878489 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d"} err="failed to get container status \"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d\": rpc error: code = NotFound desc = could not find container \"e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d\": container with ID starting with e6d65819607c1b5f8189419bb95b4c9b19ca06cb630c59aa03a5474fcce9aa3d not found: ID does not exist" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878913 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878965 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.878998 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.879093 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") pod \"07f475b7-1542-4689-b3c1-f7f4470d4a67\" (UID: \"07f475b7-1542-4689-b3c1-f7f4470d4a67\") " Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.882786 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs" (OuterVolumeSpecName: "logs") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.891713 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz" (OuterVolumeSpecName: "kube-api-access-7fwbz") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "kube-api-access-7fwbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.894700 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895159 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895176 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895195 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895202 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895219 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895225 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: E0313 16:29:30.895235 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.895241 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897382 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-metadata" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897421 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" containerName="nova-metadata-log" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897438 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" containerName="nova-cell1-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.897459 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" containerName="nova-cell0-conductor-conductor" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.898285 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.900172 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.918271 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.933473 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data" (OuterVolumeSpecName: "config-data") pod "07f475b7-1542-4689-b3c1-f7f4470d4a67" (UID: "07f475b7-1542-4689-b3c1-f7f4470d4a67"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.940132 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.956969 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.966830 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.975663 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.978676 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.980820 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981056 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981116 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x2tp\" (UniqueName: \"kubernetes.io/projected/fd1f3838-01fe-4e37-9859-8aa28048e81c-kube-api-access-6x2tp\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981197 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981295 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fwbz\" (UniqueName: \"kubernetes.io/projected/07f475b7-1542-4689-b3c1-f7f4470d4a67-kube-api-access-7fwbz\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981310 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07f475b7-1542-4689-b3c1-f7f4470d4a67-logs\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981320 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.981329 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07f475b7-1542-4689-b3c1-f7f4470d4a67-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:30 crc kubenswrapper[4907]: I0313 16:29:30.984105 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.000712 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082509 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082795 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082869 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.082905 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") pod \"b98ee93f-f32c-4199-8045-62ab95761a6a\" (UID: \"b98ee93f-f32c-4199-8045-62ab95761a6a\") " Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083375 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083408 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs" (OuterVolumeSpecName: "logs") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083429 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpn7b\" (UniqueName: \"kubernetes.io/projected/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-kube-api-access-tpn7b\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083532 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083567 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083589 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x2tp\" (UniqueName: \"kubernetes.io/projected/fd1f3838-01fe-4e37-9859-8aa28048e81c-kube-api-access-6x2tp\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.083682 4907 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98ee93f-f32c-4199-8045-62ab95761a6a-logs\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.085975 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r" (OuterVolumeSpecName: "kube-api-access-t2t6r") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "kube-api-access-t2t6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.086664 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.094518 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1f3838-01fe-4e37-9859-8aa28048e81c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.104360 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x2tp\" (UniqueName: \"kubernetes.io/projected/fd1f3838-01fe-4e37-9859-8aa28048e81c-kube-api-access-6x2tp\") pod \"nova-cell0-conductor-0\" (UID: \"fd1f3838-01fe-4e37-9859-8aa28048e81c\") " pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.114457 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.117033 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data" (OuterVolumeSpecName: "config-data") pod "b98ee93f-f32c-4199-8045-62ab95761a6a" (UID: "b98ee93f-f32c-4199-8045-62ab95761a6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185226 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpn7b\" (UniqueName: \"kubernetes.io/projected/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-kube-api-access-tpn7b\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185313 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185391 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185505 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185521 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98ee93f-f32c-4199-8045-62ab95761a6a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.185532 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2t6r\" (UniqueName: \"kubernetes.io/projected/b98ee93f-f32c-4199-8045-62ab95761a6a-kube-api-access-t2t6r\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.189079 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.190127 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.201580 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpn7b\" (UniqueName: \"kubernetes.io/projected/0a6bc64d-334f-46cd-aea9-486ab56d4cf9-kube-api-access-tpn7b\") pod \"nova-cell1-conductor-0\" (UID: \"0a6bc64d-334f-46cd-aea9-486ab56d4cf9\") " pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.330839 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.346663 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.687757 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07f475b7-1542-4689-b3c1-f7f4470d4a67","Type":"ContainerDied","Data":"55a13d8a20cd160563d8a2c5f472cc7693d88cf83045ad53675342ed80dab573"} Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.687782 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.687814 4907 scope.go:117] "RemoveContainer" containerID="047fde4a00c176dcd1e1415fa32aa7d47b1a371c3c13b528d5f7c550e42353f4" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.690592 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b98ee93f-f32c-4199-8045-62ab95761a6a","Type":"ContainerDied","Data":"6c6eae74def50447208be0f6d41599241acef5069efd81dc6d9fc3a28d2dd298"} Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.690719 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.725946 4907 scope.go:117] "RemoveContainer" containerID="c29e40d545e0c96f28dbfbc899c2ea7a2bf52224e0171c5a63ab9504fa49a30d" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.728906 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.745481 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.778079 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.784806 4907 scope.go:117] "RemoveContainer" containerID="19d030a6fb6fc190b4d45542a5b426271778e2fd40eba0953eec528508e6f63a" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.805539 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07f475b7-1542-4689-b3c1-f7f4470d4a67" path="/var/lib/kubelet/pods/07f475b7-1542-4689-b3c1-f7f4470d4a67/volumes" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.810526 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="647720e3-d21e-4f41-aeb8-d4e8f092cbb5" path="/var/lib/kubelet/pods/647720e3-d21e-4f41-aeb8-d4e8f092cbb5/volumes" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.812873 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eded65eb-7a04-4f90-ad82-be65db9014eb" path="/var/lib/kubelet/pods/eded65eb-7a04-4f90-ad82-be65db9014eb/volumes" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.813610 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.813637 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: E0313 16:29:31.813985 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.813996 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" Mar 13 16:29:31 crc kubenswrapper[4907]: E0313 16:29:31.814056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.814064 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.819009 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-api" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.819050 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" containerName="nova-api-log" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.820522 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.830128 4907 scope.go:117] "RemoveContainer" containerID="fa43cc779c11f32664399bd6d5fb18c420f2b54971c3216eb69f8689384ae861" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.830386 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.842249 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.855467 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.857426 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.862165 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.867440 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.875932 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900102 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900187 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-config-data\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900280 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13364227-6148-4b0d-ba7d-082578dfe1d5-logs\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.900338 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmdsn\" (UniqueName: \"kubernetes.io/projected/13364227-6148-4b0d-ba7d-082578dfe1d5-kube-api-access-rmdsn\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:31 crc kubenswrapper[4907]: I0313 16:29:31.934847 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.012669 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-config-data\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013132 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013194 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-logs\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013233 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-config-data\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013398 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7hj\" (UniqueName: \"kubernetes.io/projected/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-kube-api-access-zx7hj\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013446 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13364227-6148-4b0d-ba7d-082578dfe1d5-logs\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013515 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.013599 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmdsn\" (UniqueName: \"kubernetes.io/projected/13364227-6148-4b0d-ba7d-082578dfe1d5-kube-api-access-rmdsn\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.014117 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13364227-6148-4b0d-ba7d-082578dfe1d5-logs\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.018164 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.018471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13364227-6148-4b0d-ba7d-082578dfe1d5-config-data\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.029983 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmdsn\" (UniqueName: \"kubernetes.io/projected/13364227-6148-4b0d-ba7d-082578dfe1d5-kube-api-access-rmdsn\") pod \"nova-metadata-0\" (UID: \"13364227-6148-4b0d-ba7d-082578dfe1d5\") " pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.115797 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-logs\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.115952 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7hj\" (UniqueName: \"kubernetes.io/projected/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-kube-api-access-zx7hj\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.115996 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.116070 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-config-data\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.116738 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-logs\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.120148 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.129445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-config-data\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.131181 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7hj\" (UniqueName: \"kubernetes.io/projected/c11f341d-9a76-4e7e-b1e7-8d63c30bac32-kube-api-access-zx7hj\") pod \"nova-api-0\" (UID: \"c11f341d-9a76-4e7e-b1e7-8d63c30bac32\") " pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.142815 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.221595 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.584146 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 16:29:32 crc kubenswrapper[4907]: W0313 16:29:32.586856 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13364227_6148_4b0d_ba7d_082578dfe1d5.slice/crio-1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083 WatchSource:0}: Error finding container 1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083: Status 404 returned error can't find the container with id 1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083 Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.605351 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.711268 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13364227-6148-4b0d-ba7d-082578dfe1d5","Type":"ContainerStarted","Data":"1ff2b23316c389fffd85d9082d46b0eb2f2425c20a257e387b3147bcbd87d083"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.713047 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fd1f3838-01fe-4e37-9859-8aa28048e81c","Type":"ContainerStarted","Data":"2a9b247acff02796096f1440a3319cf378894bfe822c7a2d8316e4926ccfc23a"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.713074 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"fd1f3838-01fe-4e37-9859-8aa28048e81c","Type":"ContainerStarted","Data":"fe9e322f499f6530211e73b3ccc8acf23f16f152f1f35fe4ed828c9adb388d70"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.713399 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716023 4907 generic.go:334] "Generic (PLEG): container finished" podID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" exitCode=0 Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716135 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerDied","Data":"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716344 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"48adeb28-f850-4e7c-a546-2c0b32d1b16c","Type":"ContainerDied","Data":"bba3814d9581f3c03937b4a23ce7f341cbe05c058698e236d16b4dfa33146288"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.716362 4907 scope.go:117] "RemoveContainer" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.719715 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0a6bc64d-334f-46cd-aea9-486ab56d4cf9","Type":"ContainerStarted","Data":"a74f56ca47951bf2450e7cf3e0fb341611dfcfee373d5cc75aa42a2e60567167"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.719758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0a6bc64d-334f-46cd-aea9-486ab56d4cf9","Type":"ContainerStarted","Data":"32bb38cda79c6777a5df03efdb4d4b0b1b79e95ac6d0ad101b98b5d9b64830ef"} Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.720182 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.731915 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") pod \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.732282 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") pod \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.732547 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") pod \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\" (UID: \"48adeb28-f850-4e7c-a546-2c0b32d1b16c\") " Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.738159 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq" (OuterVolumeSpecName: "kube-api-access-bm6cq") pod "48adeb28-f850-4e7c-a546-2c0b32d1b16c" (UID: "48adeb28-f850-4e7c-a546-2c0b32d1b16c"). InnerVolumeSpecName "kube-api-access-bm6cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.758499 4907 scope.go:117] "RemoveContainer" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" Mar 13 16:29:32 crc kubenswrapper[4907]: E0313 16:29:32.759612 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88\": container with ID starting with 5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88 not found: ID does not exist" containerID="5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.759667 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88"} err="failed to get container status \"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88\": rpc error: code = NotFound desc = could not find container \"5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88\": container with ID starting with 5de9510392b1c0241095d60bc966293a382f722cf2d981565b650d57c091ce88 not found: ID does not exist" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.766655 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.766636185 podStartE2EDuration="2.766636185s" podCreationTimestamp="2026-03-13 16:29:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:32.728978856 +0000 UTC m=+8671.628766555" watchObservedRunningTime="2026-03-13 16:29:32.766636185 +0000 UTC m=+8671.666423874" Mar 13 16:29:32 crc kubenswrapper[4907]: W0313 16:29:32.773641 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc11f341d_9a76_4e7e_b1e7_8d63c30bac32.slice/crio-0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194 WatchSource:0}: Error finding container 0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194: Status 404 returned error can't find the container with id 0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194 Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.774379 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.774359343 podStartE2EDuration="2.774359343s" podCreationTimestamp="2026-03-13 16:29:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:32.742803525 +0000 UTC m=+8671.642591214" watchObservedRunningTime="2026-03-13 16:29:32.774359343 +0000 UTC m=+8671.674147032" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.775938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data" (OuterVolumeSpecName: "config-data") pod "48adeb28-f850-4e7c-a546-2c0b32d1b16c" (UID: "48adeb28-f850-4e7c-a546-2c0b32d1b16c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.784176 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:32 crc kubenswrapper[4907]: E0313 16:29:32.784379 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.798005 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.805354 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48adeb28-f850-4e7c-a546-2c0b32d1b16c" (UID: "48adeb28-f850-4e7c-a546-2c0b32d1b16c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.835521 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm6cq\" (UniqueName: \"kubernetes.io/projected/48adeb28-f850-4e7c-a546-2c0b32d1b16c-kube-api-access-bm6cq\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.835712 4907 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:32 crc kubenswrapper[4907]: I0313 16:29:32.835771 4907 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48adeb28-f850-4e7c-a546-2c0b32d1b16c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.074229 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.084865 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.131921 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: E0313 16:29:33.132676 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.132695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.133019 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" containerName="nova-scheduler-scheduler" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.134053 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.137352 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.148269 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.252448 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7kh\" (UniqueName: \"kubernetes.io/projected/54123701-97fb-4f4f-baed-4c2183bad203-kube-api-access-bv7kh\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.252526 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.252666 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-config-data\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.354722 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-config-data\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.355133 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7kh\" (UniqueName: \"kubernetes.io/projected/54123701-97fb-4f4f-baed-4c2183bad203-kube-api-access-bv7kh\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.355213 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.362635 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-config-data\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.365471 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54123701-97fb-4f4f-baed-4c2183bad203-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.377107 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7kh\" (UniqueName: \"kubernetes.io/projected/54123701-97fb-4f4f-baed-4c2183bad203-kube-api-access-bv7kh\") pod \"nova-scheduler-0\" (UID: \"54123701-97fb-4f4f-baed-4c2183bad203\") " pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.482212 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.729950 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c11f341d-9a76-4e7e-b1e7-8d63c30bac32","Type":"ContainerStarted","Data":"961fd42c70e8f82176c365ec5a810b2861304a70794684f34381af8052b330e0"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.730326 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c11f341d-9a76-4e7e-b1e7-8d63c30bac32","Type":"ContainerStarted","Data":"f5d734a1b9cc16fa2be7da399d5b434256b5ef06f84511b5c4f404f76cbf3a3d"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.730339 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c11f341d-9a76-4e7e-b1e7-8d63c30bac32","Type":"ContainerStarted","Data":"0a9826a17473ee42d705880efe1f102dabc505ebf63c497c5a550f34dd34a194"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.736465 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13364227-6148-4b0d-ba7d-082578dfe1d5","Type":"ContainerStarted","Data":"0ced5e91ca7c08e6c13f6c1fcd52b7cec01a06c9dc08fd56db8b89c560a582f0"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.736491 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"13364227-6148-4b0d-ba7d-082578dfe1d5","Type":"ContainerStarted","Data":"8f1587fc1344b2989add0ef318fec5bbb239ecb1b3e8ef09fe25211d417e344b"} Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.749917 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.749898742 podStartE2EDuration="2.749898742s" podCreationTimestamp="2026-03-13 16:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:33.745199179 +0000 UTC m=+8672.644986888" watchObservedRunningTime="2026-03-13 16:29:33.749898742 +0000 UTC m=+8672.649686441" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.785826 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.785809831 podStartE2EDuration="2.785809831s" podCreationTimestamp="2026-03-13 16:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:33.768611737 +0000 UTC m=+8672.668399426" watchObservedRunningTime="2026-03-13 16:29:33.785809831 +0000 UTC m=+8672.685597520" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.793712 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48adeb28-f850-4e7c-a546-2c0b32d1b16c" path="/var/lib/kubelet/pods/48adeb28-f850-4e7c-a546-2c0b32d1b16c/volumes" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.794452 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b98ee93f-f32c-4199-8045-62ab95761a6a" path="/var/lib/kubelet/pods/b98ee93f-f32c-4199-8045-62ab95761a6a/volumes" Mar 13 16:29:33 crc kubenswrapper[4907]: I0313 16:29:33.949443 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 16:29:33 crc kubenswrapper[4907]: W0313 16:29:33.960449 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54123701_97fb_4f4f_baed_4c2183bad203.slice/crio-c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00 WatchSource:0}: Error finding container c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00: Status 404 returned error can't find the container with id c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00 Mar 13 16:29:34 crc kubenswrapper[4907]: I0313 16:29:34.750420 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"54123701-97fb-4f4f-baed-4c2183bad203","Type":"ContainerStarted","Data":"a4b11035cfa2df06f9b86ebf95edee9ab566cdaa742b9648bb53b6e9192440e2"} Mar 13 16:29:34 crc kubenswrapper[4907]: I0313 16:29:34.750726 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"54123701-97fb-4f4f-baed-4c2183bad203","Type":"ContainerStarted","Data":"c5374a10aaafdfc348fb5411307fafd2b204901548a49c65dbd3dfb49e078a00"} Mar 13 16:29:34 crc kubenswrapper[4907]: I0313 16:29:34.787759 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.782571156 podStartE2EDuration="1.782571156s" podCreationTimestamp="2026-03-13 16:29:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 16:29:34.775043595 +0000 UTC m=+8673.674831354" watchObservedRunningTime="2026-03-13 16:29:34.782571156 +0000 UTC m=+8673.682358885" Mar 13 16:29:38 crc kubenswrapper[4907]: I0313 16:29:38.483627 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 16:29:41 crc kubenswrapper[4907]: I0313 16:29:41.395762 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 16:29:41 crc kubenswrapper[4907]: I0313 16:29:41.410962 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.145032 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.145403 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.222679 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 16:29:42 crc kubenswrapper[4907]: I0313 16:29:42.222756 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.228373 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="13364227-6148-4b0d-ba7d-082578dfe1d5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.0.8:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.228453 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="13364227-6148-4b0d-ba7d-082578dfe1d5" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.0.8:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.310076 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c11f341d-9a76-4e7e-b1e7-8d63c30bac32" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.11:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.310095 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c11f341d-9a76-4e7e-b1e7-8d63c30bac32" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.11:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.483817 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.524802 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 16:29:43 crc kubenswrapper[4907]: I0313 16:29:43.876676 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 16:29:45 crc kubenswrapper[4907]: I0313 16:29:45.783669 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:45 crc kubenswrapper[4907]: E0313 16:29:45.784267 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.143114 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.144646 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.222651 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 16:29:50 crc kubenswrapper[4907]: I0313 16:29:50.222715 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.146060 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.146469 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.150015 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.152678 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.230323 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.233902 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.243172 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 16:29:52 crc kubenswrapper[4907]: I0313 16:29:52.962055 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 16:29:59 crc kubenswrapper[4907]: I0313 16:29:59.783162 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:29:59 crc kubenswrapper[4907]: E0313 16:29:59.784263 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.151049 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.153137 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.155109 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.156871 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.164947 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.167734 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.169603 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.169950 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.177980 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.178984 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.192485 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341544 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"auto-csr-approver-29556990-4jrg4\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341628 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.341734 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443768 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"auto-csr-approver-29556990-4jrg4\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443854 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443901 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.443964 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.444915 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.471917 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.474753 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"auto-csr-approver-29556990-4jrg4\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.475651 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"collect-profiles-29556990-mgpgm\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.501167 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.511837 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:00 crc kubenswrapper[4907]: I0313 16:30:00.964338 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm"] Mar 13 16:30:01 crc kubenswrapper[4907]: I0313 16:30:01.056538 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" event={"ID":"eb5ba94e-890e-4791-b2f6-e30d1d92ed96","Type":"ContainerStarted","Data":"88f370a6bce1c1e14ee860380282bda5a2ea2cf44f416fcec26c54a386636978"} Mar 13 16:30:01 crc kubenswrapper[4907]: W0313 16:30:01.081283 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a413954_e701_4cec_8868_4fcdaecfabe8.slice/crio-7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71 WatchSource:0}: Error finding container 7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71: Status 404 returned error can't find the container with id 7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71 Mar 13 16:30:01 crc kubenswrapper[4907]: I0313 16:30:01.086726 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:30:02 crc kubenswrapper[4907]: I0313 16:30:02.074316 4907 generic.go:334] "Generic (PLEG): container finished" podID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerID="403679ef0fad99e187e6956860663a7b80495194144d0c59af9ade6f000f4fe6" exitCode=0 Mar 13 16:30:02 crc kubenswrapper[4907]: I0313 16:30:02.074403 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" event={"ID":"eb5ba94e-890e-4791-b2f6-e30d1d92ed96","Type":"ContainerDied","Data":"403679ef0fad99e187e6956860663a7b80495194144d0c59af9ade6f000f4fe6"} Mar 13 16:30:02 crc kubenswrapper[4907]: I0313 16:30:02.078136 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" event={"ID":"5a413954-e701-4cec-8868-4fcdaecfabe8","Type":"ContainerStarted","Data":"7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71"} Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.484830 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.617209 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") pod \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.617262 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") pod \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.617307 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") pod \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\" (UID: \"eb5ba94e-890e-4791-b2f6-e30d1d92ed96\") " Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.618531 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume" (OuterVolumeSpecName: "config-volume") pod "eb5ba94e-890e-4791-b2f6-e30d1d92ed96" (UID: "eb5ba94e-890e-4791-b2f6-e30d1d92ed96"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.626503 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t" (OuterVolumeSpecName: "kube-api-access-n5k8t") pod "eb5ba94e-890e-4791-b2f6-e30d1d92ed96" (UID: "eb5ba94e-890e-4791-b2f6-e30d1d92ed96"). InnerVolumeSpecName "kube-api-access-n5k8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.626526 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eb5ba94e-890e-4791-b2f6-e30d1d92ed96" (UID: "eb5ba94e-890e-4791-b2f6-e30d1d92ed96"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.719996 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.720028 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5k8t\" (UniqueName: \"kubernetes.io/projected/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-kube-api-access-n5k8t\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:03 crc kubenswrapper[4907]: I0313 16:30:03.720040 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eb5ba94e-890e-4791-b2f6-e30d1d92ed96-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.102276 4907 generic.go:334] "Generic (PLEG): container finished" podID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerID="5585587b3788c101f53b09ee74a77a65c4d4a4773da4d3e62432619e4599efa1" exitCode=0 Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.102392 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" event={"ID":"5a413954-e701-4cec-8868-4fcdaecfabe8","Type":"ContainerDied","Data":"5585587b3788c101f53b09ee74a77a65c4d4a4773da4d3e62432619e4599efa1"} Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.106565 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" event={"ID":"eb5ba94e-890e-4791-b2f6-e30d1d92ed96","Type":"ContainerDied","Data":"88f370a6bce1c1e14ee860380282bda5a2ea2cf44f416fcec26c54a386636978"} Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.106613 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88f370a6bce1c1e14ee860380282bda5a2ea2cf44f416fcec26c54a386636978" Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.106630 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556990-mgpgm" Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.608007 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 16:30:04 crc kubenswrapper[4907]: I0313 16:30:04.622448 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556945-vp2lp"] Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.575873 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.666767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") pod \"5a413954-e701-4cec-8868-4fcdaecfabe8\" (UID: \"5a413954-e701-4cec-8868-4fcdaecfabe8\") " Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.675135 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9" (OuterVolumeSpecName: "kube-api-access-86rp9") pod "5a413954-e701-4cec-8868-4fcdaecfabe8" (UID: "5a413954-e701-4cec-8868-4fcdaecfabe8"). InnerVolumeSpecName "kube-api-access-86rp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.769161 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86rp9\" (UniqueName: \"kubernetes.io/projected/5a413954-e701-4cec-8868-4fcdaecfabe8-kube-api-access-86rp9\") on node \"crc\" DevicePath \"\"" Mar 13 16:30:05 crc kubenswrapper[4907]: I0313 16:30:05.796554 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="329cddab-6958-4dbb-a2d0-baef50885413" path="/var/lib/kubelet/pods/329cddab-6958-4dbb-a2d0-baef50885413/volumes" Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.131087 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" event={"ID":"5a413954-e701-4cec-8868-4fcdaecfabe8","Type":"ContainerDied","Data":"7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71"} Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.131138 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556990-4jrg4" Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.131152 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7efa7e899bdff3abb97dc670393bea14cc20267410e92c626acaeba445e97a71" Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.625504 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:30:06 crc kubenswrapper[4907]: I0313 16:30:06.633944 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556984-2lfjr"] Mar 13 16:30:07 crc kubenswrapper[4907]: I0313 16:30:07.797451 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0090a6a-15d8-4445-9830-9cab1aec7fdb" path="/var/lib/kubelet/pods/d0090a6a-15d8-4445-9830-9cab1aec7fdb/volumes" Mar 13 16:30:10 crc kubenswrapper[4907]: I0313 16:30:10.783680 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:10 crc kubenswrapper[4907]: E0313 16:30:10.784626 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:25 crc kubenswrapper[4907]: I0313 16:30:25.783502 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:25 crc kubenswrapper[4907]: E0313 16:30:25.784568 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:36 crc kubenswrapper[4907]: I0313 16:30:36.782854 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:36 crc kubenswrapper[4907]: E0313 16:30:36.784185 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:30:43 crc kubenswrapper[4907]: I0313 16:30:43.496183 4907 scope.go:117] "RemoveContainer" containerID="65a8dfeb436104f422bc6ebeae3f1c135d2d01201103445d064625f2317603ad" Mar 13 16:30:43 crc kubenswrapper[4907]: I0313 16:30:43.567577 4907 scope.go:117] "RemoveContainer" containerID="5fb59125e06d01b85c52baad33f05fd7f12c902ae1f36e7af1e62fdbcb0056b5" Mar 13 16:30:51 crc kubenswrapper[4907]: I0313 16:30:51.802068 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:30:51 crc kubenswrapper[4907]: E0313 16:30:51.803387 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:06 crc kubenswrapper[4907]: I0313 16:31:06.783062 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:06 crc kubenswrapper[4907]: E0313 16:31:06.784326 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:21 crc kubenswrapper[4907]: I0313 16:31:21.796193 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:21 crc kubenswrapper[4907]: E0313 16:31:21.797018 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:33 crc kubenswrapper[4907]: I0313 16:31:33.782874 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:33 crc kubenswrapper[4907]: E0313 16:31:33.783849 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:48 crc kubenswrapper[4907]: I0313 16:31:48.782815 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:48 crc kubenswrapper[4907]: E0313 16:31:48.783609 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.105821 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:31:49 crc kubenswrapper[4907]: E0313 16:31:49.106575 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerName="oc" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106601 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerName="oc" Mar 13 16:31:49 crc kubenswrapper[4907]: E0313 16:31:49.106632 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerName="collect-profiles" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106641 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerName="collect-profiles" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106933 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb5ba94e-890e-4791-b2f6-e30d1d92ed96" containerName="collect-profiles" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.106968 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" containerName="oc" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.108920 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.121622 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.159552 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.159769 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.159956 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.262430 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.262534 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.262608 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.263113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.263538 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.295802 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"redhat-operators-bfz29\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.429108 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:49 crc kubenswrapper[4907]: I0313 16:31:49.967756 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:31:50 crc kubenswrapper[4907]: I0313 16:31:50.272806 4907 generic.go:334] "Generic (PLEG): container finished" podID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" exitCode=0 Mar 13 16:31:50 crc kubenswrapper[4907]: I0313 16:31:50.272858 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c"} Mar 13 16:31:50 crc kubenswrapper[4907]: I0313 16:31:50.273105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerStarted","Data":"6843dee463e8aa2139eea25e2bbbae5a7e3cae3bb8c1d0992495315ec9382157"} Mar 13 16:31:52 crc kubenswrapper[4907]: I0313 16:31:52.294818 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerStarted","Data":"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df"} Mar 13 16:31:55 crc kubenswrapper[4907]: I0313 16:31:55.325789 4907 generic.go:334] "Generic (PLEG): container finished" podID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" exitCode=0 Mar 13 16:31:55 crc kubenswrapper[4907]: I0313 16:31:55.325854 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df"} Mar 13 16:31:56 crc kubenswrapper[4907]: I0313 16:31:56.341446 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerStarted","Data":"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410"} Mar 13 16:31:56 crc kubenswrapper[4907]: I0313 16:31:56.379839 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bfz29" podStartSLOduration=1.901810102 podStartE2EDuration="7.379818269s" podCreationTimestamp="2026-03-13 16:31:49 +0000 UTC" firstStartedPulling="2026-03-13 16:31:50.274424365 +0000 UTC m=+8809.174212054" lastFinishedPulling="2026-03-13 16:31:55.752432532 +0000 UTC m=+8814.652220221" observedRunningTime="2026-03-13 16:31:56.366947947 +0000 UTC m=+8815.266735686" watchObservedRunningTime="2026-03-13 16:31:56.379818269 +0000 UTC m=+8815.279605968" Mar 13 16:31:59 crc kubenswrapper[4907]: I0313 16:31:59.430352 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:59 crc kubenswrapper[4907]: I0313 16:31:59.430797 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:31:59 crc kubenswrapper[4907]: I0313 16:31:59.783794 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:31:59 crc kubenswrapper[4907]: E0313 16:31:59.784396 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.145605 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.147008 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.154977 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.155071 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.155302 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.159648 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.209644 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"auto-csr-approver-29556992-qj2mm\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.311495 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"auto-csr-approver-29556992-qj2mm\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.338747 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"auto-csr-approver-29556992-qj2mm\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.468261 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.490244 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bfz29" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" probeResult="failure" output=< Mar 13 16:32:00 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:32:00 crc kubenswrapper[4907]: > Mar 13 16:32:00 crc kubenswrapper[4907]: I0313 16:32:00.969924 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:32:01 crc kubenswrapper[4907]: I0313 16:32:01.416036 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerStarted","Data":"8b6a97332a7098fe96c5bc5d67ef2f10bf6399130d709ccb2cc03a0dd3fd7fc1"} Mar 13 16:32:02 crc kubenswrapper[4907]: I0313 16:32:02.426758 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerStarted","Data":"a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a"} Mar 13 16:32:02 crc kubenswrapper[4907]: I0313 16:32:02.443466 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" podStartSLOduration=1.470477152 podStartE2EDuration="2.443449568s" podCreationTimestamp="2026-03-13 16:32:00 +0000 UTC" firstStartedPulling="2026-03-13 16:32:00.955284381 +0000 UTC m=+8819.855072060" lastFinishedPulling="2026-03-13 16:32:01.928256797 +0000 UTC m=+8820.828044476" observedRunningTime="2026-03-13 16:32:02.439464046 +0000 UTC m=+8821.339251745" watchObservedRunningTime="2026-03-13 16:32:02.443449568 +0000 UTC m=+8821.343237257" Mar 13 16:32:03 crc kubenswrapper[4907]: I0313 16:32:03.441606 4907 generic.go:334] "Generic (PLEG): container finished" podID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerID="a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a" exitCode=0 Mar 13 16:32:03 crc kubenswrapper[4907]: I0313 16:32:03.441719 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerDied","Data":"a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a"} Mar 13 16:32:04 crc kubenswrapper[4907]: I0313 16:32:04.897512 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:04 crc kubenswrapper[4907]: I0313 16:32:04.924843 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") pod \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\" (UID: \"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476\") " Mar 13 16:32:04 crc kubenswrapper[4907]: I0313 16:32:04.934124 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6" (OuterVolumeSpecName: "kube-api-access-d55s6") pod "5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" (UID: "5d2dfd3f-0bc6-48f9-85f7-100ef8f23476"). InnerVolumeSpecName "kube-api-access-d55s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.028230 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d55s6\" (UniqueName: \"kubernetes.io/projected/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476-kube-api-access-d55s6\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.471334 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" event={"ID":"5d2dfd3f-0bc6-48f9-85f7-100ef8f23476","Type":"ContainerDied","Data":"8b6a97332a7098fe96c5bc5d67ef2f10bf6399130d709ccb2cc03a0dd3fd7fc1"} Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.471380 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b6a97332a7098fe96c5bc5d67ef2f10bf6399130d709ccb2cc03a0dd3fd7fc1" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.471483 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556992-qj2mm" Mar 13 16:32:05 crc kubenswrapper[4907]: I0313 16:32:05.996536 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:32:06 crc kubenswrapper[4907]: I0313 16:32:06.028155 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556986-nlfck"] Mar 13 16:32:07 crc kubenswrapper[4907]: I0313 16:32:07.799025 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4" path="/var/lib/kubelet/pods/1bfa4f11-b5c4-4767-bb8e-3940faa2a5f4/volumes" Mar 13 16:32:09 crc kubenswrapper[4907]: I0313 16:32:09.511270 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:09 crc kubenswrapper[4907]: I0313 16:32:09.577505 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:09 crc kubenswrapper[4907]: I0313 16:32:09.754097 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:32:11 crc kubenswrapper[4907]: I0313 16:32:11.545779 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bfz29" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" containerID="cri-o://9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" gracePeriod=2 Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.243094 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.397375 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") pod \"fed253af-6b1e-49b2-9d21-4467f33ecf94\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.397639 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") pod \"fed253af-6b1e-49b2-9d21-4467f33ecf94\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.397855 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") pod \"fed253af-6b1e-49b2-9d21-4467f33ecf94\" (UID: \"fed253af-6b1e-49b2-9d21-4467f33ecf94\") " Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.399243 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities" (OuterVolumeSpecName: "utilities") pod "fed253af-6b1e-49b2-9d21-4467f33ecf94" (UID: "fed253af-6b1e-49b2-9d21-4467f33ecf94"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.405808 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8" (OuterVolumeSpecName: "kube-api-access-xmch8") pod "fed253af-6b1e-49b2-9d21-4467f33ecf94" (UID: "fed253af-6b1e-49b2-9d21-4467f33ecf94"). InnerVolumeSpecName "kube-api-access-xmch8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.501593 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.501635 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmch8\" (UniqueName: \"kubernetes.io/projected/fed253af-6b1e-49b2-9d21-4467f33ecf94-kube-api-access-xmch8\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561330 4907 generic.go:334] "Generic (PLEG): container finished" podID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" exitCode=0 Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561377 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410"} Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561416 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bfz29" event={"ID":"fed253af-6b1e-49b2-9d21-4467f33ecf94","Type":"ContainerDied","Data":"6843dee463e8aa2139eea25e2bbbae5a7e3cae3bb8c1d0992495315ec9382157"} Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561437 4907 scope.go:117] "RemoveContainer" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.561632 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bfz29" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.571282 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fed253af-6b1e-49b2-9d21-4467f33ecf94" (UID: "fed253af-6b1e-49b2-9d21-4467f33ecf94"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.597505 4907 scope.go:117] "RemoveContainer" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.606273 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fed253af-6b1e-49b2-9d21-4467f33ecf94-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.634012 4907 scope.go:117] "RemoveContainer" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.709541 4907 scope.go:117] "RemoveContainer" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" Mar 13 16:32:12 crc kubenswrapper[4907]: E0313 16:32:12.711511 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410\": container with ID starting with 9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410 not found: ID does not exist" containerID="9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.711607 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410"} err="failed to get container status \"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410\": rpc error: code = NotFound desc = could not find container \"9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410\": container with ID starting with 9064e0377ba2c4d96cb954df5a4d9d8e4208784b7270f9410b0d253f7b5f4410 not found: ID does not exist" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.711644 4907 scope.go:117] "RemoveContainer" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" Mar 13 16:32:12 crc kubenswrapper[4907]: E0313 16:32:12.712232 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df\": container with ID starting with b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df not found: ID does not exist" containerID="b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.712266 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df"} err="failed to get container status \"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df\": rpc error: code = NotFound desc = could not find container \"b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df\": container with ID starting with b100747e5d6556c4ee784efb31b617f6c9c90e6075eb0a8d0df0270a93d3d2df not found: ID does not exist" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.712286 4907 scope.go:117] "RemoveContainer" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" Mar 13 16:32:12 crc kubenswrapper[4907]: E0313 16:32:12.714123 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c\": container with ID starting with 80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c not found: ID does not exist" containerID="80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.714230 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c"} err="failed to get container status \"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c\": rpc error: code = NotFound desc = could not find container \"80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c\": container with ID starting with 80c7c01d67fbd09368a3d7aec0c7facbc1ce68808d7cf94f486937495c9b6a4c not found: ID does not exist" Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.899500 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:32:12 crc kubenswrapper[4907]: I0313 16:32:12.909130 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bfz29"] Mar 13 16:32:13 crc kubenswrapper[4907]: I0313 16:32:13.796497 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" path="/var/lib/kubelet/pods/fed253af-6b1e-49b2-9d21-4467f33ecf94/volumes" Mar 13 16:32:14 crc kubenswrapper[4907]: I0313 16:32:14.782666 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:32:14 crc kubenswrapper[4907]: E0313 16:32:14.783260 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:32:26 crc kubenswrapper[4907]: I0313 16:32:26.783753 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:32:27 crc kubenswrapper[4907]: I0313 16:32:27.749911 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9"} Mar 13 16:32:43 crc kubenswrapper[4907]: I0313 16:32:43.676293 4907 scope.go:117] "RemoveContainer" containerID="4fc188ddf095682c93e7e1f3f3249b322a5fe91043eda372761cb774c37d43db" Mar 13 16:33:08 crc kubenswrapper[4907]: I0313 16:33:08.223930 4907 generic.go:334] "Generic (PLEG): container finished" podID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerID="8644ffba95037a0d279e1ad31f88e2f9c3fae28fb5fd51859c1fb3e25cec04cc" exitCode=0 Mar 13 16:33:08 crc kubenswrapper[4907]: I0313 16:33:08.224160 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerDied","Data":"8644ffba95037a0d279e1ad31f88e2f9c3fae28fb5fd51859c1fb3e25cec04cc"} Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.706482 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841592 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841653 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841754 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841805 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.841937 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842009 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842047 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842086 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842205 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842259 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842300 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842358 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.842447 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") pod \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\" (UID: \"de68abba-24e7-49b8-9e1f-adfa05d9d3dd\") " Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.848446 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph" (OuterVolumeSpecName: "ceph") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.849349 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g" (OuterVolumeSpecName: "kube-api-access-njn9g") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "kube-api-access-njn9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.852436 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.873136 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.875728 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.894251 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.895769 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.896132 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory" (OuterVolumeSpecName: "inventory") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.897933 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.904672 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.907860 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.909369 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.918096 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "de68abba-24e7-49b8-9e1f-adfa05d9d3dd" (UID: "de68abba-24e7-49b8-9e1f-adfa05d9d3dd"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945662 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njn9g\" (UniqueName: \"kubernetes.io/projected/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-kube-api-access-njn9g\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945709 4907 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ceph\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945730 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945748 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945765 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945783 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945798 4907 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945813 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945827 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945843 4907 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945858 4907 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945872 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:09 crc kubenswrapper[4907]: I0313 16:33:09.945911 4907 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de68abba-24e7-49b8-9e1f-adfa05d9d3dd-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 16:33:10 crc kubenswrapper[4907]: I0313 16:33:10.249105 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" event={"ID":"de68abba-24e7-49b8-9e1f-adfa05d9d3dd","Type":"ContainerDied","Data":"315fdd9184d5c2a2780c5ff803490daf8312afa1aa0a6d15f69dbc1f771bf80c"} Mar 13 16:33:10 crc kubenswrapper[4907]: I0313 16:33:10.249390 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="315fdd9184d5c2a2780c5ff803490daf8312afa1aa0a6d15f69dbc1f771bf80c" Mar 13 16:33:10 crc kubenswrapper[4907]: I0313 16:33:10.249273 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc" Mar 13 16:33:23 crc kubenswrapper[4907]: E0313 16:33:23.493449 4907 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.174:33374->38.102.83.174:33349: write tcp 38.102.83.174:33374->38.102.83.174:33349: write: broken pipe Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.188188 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.189993 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-utilities" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190022 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-utilities" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190053 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerName="oc" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190061 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerName="oc" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190100 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190150 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190198 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190207 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" Mar 13 16:34:00 crc kubenswrapper[4907]: E0313 16:34:00.190217 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-content" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190224 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="extract-content" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190829 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="fed253af-6b1e-49b2-9d21-4467f33ecf94" containerName="registry-server" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190901 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" containerName="oc" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.190932 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="de68abba-24e7-49b8-9e1f-adfa05d9d3dd" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.192629 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.196703 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.197449 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.197645 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.237573 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.304393 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"auto-csr-approver-29556994-452h2\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.407513 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"auto-csr-approver-29556994-452h2\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.441703 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"auto-csr-approver-29556994-452h2\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:00 crc kubenswrapper[4907]: I0313 16:34:00.520630 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:01 crc kubenswrapper[4907]: I0313 16:34:01.076434 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:34:01 crc kubenswrapper[4907]: I0313 16:34:01.091525 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:34:01 crc kubenswrapper[4907]: I0313 16:34:01.919433 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556994-452h2" event={"ID":"262afeb8-16de-47e7-a64b-5b4e938aa1f4","Type":"ContainerStarted","Data":"a8d9553ed7e2900c7f96d66573bd4bda4a8ba968c5d8cee4a3dbeb326577c98c"} Mar 13 16:34:02 crc kubenswrapper[4907]: I0313 16:34:02.937052 4907 generic.go:334] "Generic (PLEG): container finished" podID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerID="237bb05fcc8f490c8ced9263406f8715e053a689a6c929dacb3684e153ea9c27" exitCode=0 Mar 13 16:34:02 crc kubenswrapper[4907]: I0313 16:34:02.937147 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556994-452h2" event={"ID":"262afeb8-16de-47e7-a64b-5b4e938aa1f4","Type":"ContainerDied","Data":"237bb05fcc8f490c8ced9263406f8715e053a689a6c929dacb3684e153ea9c27"} Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.424259 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.608018 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") pod \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\" (UID: \"262afeb8-16de-47e7-a64b-5b4e938aa1f4\") " Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.617725 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz" (OuterVolumeSpecName: "kube-api-access-wqmsz") pod "262afeb8-16de-47e7-a64b-5b4e938aa1f4" (UID: "262afeb8-16de-47e7-a64b-5b4e938aa1f4"). InnerVolumeSpecName "kube-api-access-wqmsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.710914 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wqmsz\" (UniqueName: \"kubernetes.io/projected/262afeb8-16de-47e7-a64b-5b4e938aa1f4-kube-api-access-wqmsz\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.969828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556994-452h2" event={"ID":"262afeb8-16de-47e7-a64b-5b4e938aa1f4","Type":"ContainerDied","Data":"a8d9553ed7e2900c7f96d66573bd4bda4a8ba968c5d8cee4a3dbeb326577c98c"} Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.970382 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8d9553ed7e2900c7f96d66573bd4bda4a8ba968c5d8cee4a3dbeb326577c98c" Mar 13 16:34:04 crc kubenswrapper[4907]: I0313 16:34:04.970469 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556994-452h2" Mar 13 16:34:05 crc kubenswrapper[4907]: I0313 16:34:05.545198 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:34:05 crc kubenswrapper[4907]: I0313 16:34:05.558753 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556988-t5tgt"] Mar 13 16:34:05 crc kubenswrapper[4907]: I0313 16:34:05.799321 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa9f963-9a2b-4101-88e5-9553783a9963" path="/var/lib/kubelet/pods/6aa9f963-9a2b-4101-88e5-9553783a9963/volumes" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.862423 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:11 crc kubenswrapper[4907]: E0313 16:34:11.863738 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerName="oc" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.863754 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerName="oc" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.864087 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" containerName="oc" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.865750 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:11 crc kubenswrapper[4907]: I0313 16:34:11.888479 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.039499 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.039546 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.039565 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.142484 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.142540 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.142564 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.143170 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.143237 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.167979 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"certified-operators-598sp\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.200299 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:12 crc kubenswrapper[4907]: I0313 16:34:12.758702 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:13 crc kubenswrapper[4907]: I0313 16:34:13.068137 4907 generic.go:334] "Generic (PLEG): container finished" podID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" exitCode=0 Mar 13 16:34:13 crc kubenswrapper[4907]: I0313 16:34:13.068235 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a"} Mar 13 16:34:13 crc kubenswrapper[4907]: I0313 16:34:13.068527 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerStarted","Data":"87a43c517510df3cce3a23c229d4e56247da0404420b2f98fa527a9b52aeadfb"} Mar 13 16:34:15 crc kubenswrapper[4907]: I0313 16:34:15.099511 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerStarted","Data":"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4"} Mar 13 16:34:16 crc kubenswrapper[4907]: I0313 16:34:16.121600 4907 generic.go:334] "Generic (PLEG): container finished" podID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" exitCode=0 Mar 13 16:34:16 crc kubenswrapper[4907]: I0313 16:34:16.122018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4"} Mar 13 16:34:17 crc kubenswrapper[4907]: I0313 16:34:17.134610 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerStarted","Data":"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880"} Mar 13 16:34:17 crc kubenswrapper[4907]: I0313 16:34:17.158586 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-598sp" podStartSLOduration=2.685575957 podStartE2EDuration="6.15856231s" podCreationTimestamp="2026-03-13 16:34:11 +0000 UTC" firstStartedPulling="2026-03-13 16:34:13.0706122 +0000 UTC m=+8951.970399889" lastFinishedPulling="2026-03-13 16:34:16.543598553 +0000 UTC m=+8955.443386242" observedRunningTime="2026-03-13 16:34:17.151409468 +0000 UTC m=+8956.051197157" watchObservedRunningTime="2026-03-13 16:34:17.15856231 +0000 UTC m=+8956.058349999" Mar 13 16:34:22 crc kubenswrapper[4907]: I0313 16:34:22.200576 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:22 crc kubenswrapper[4907]: I0313 16:34:22.201274 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:22 crc kubenswrapper[4907]: I0313 16:34:22.888035 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:23 crc kubenswrapper[4907]: I0313 16:34:23.298689 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:23 crc kubenswrapper[4907]: I0313 16:34:23.354337 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.257841 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-598sp" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" containerID="cri-o://61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" gracePeriod=2 Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.561966 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.564875 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.578274 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.748459 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.748652 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.749135 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.824464 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.851125 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.851491 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.851597 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.852225 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.853399 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.876861 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"community-operators-kksrr\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.894494 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.954166 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") pod \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.954236 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") pod \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.954328 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") pod \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\" (UID: \"92a3a8bb-d3df-4489-8cf5-aacdad7516a9\") " Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.959476 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities" (OuterVolumeSpecName: "utilities") pod "92a3a8bb-d3df-4489-8cf5-aacdad7516a9" (UID: "92a3a8bb-d3df-4489-8cf5-aacdad7516a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:25 crc kubenswrapper[4907]: I0313 16:34:25.969655 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq" (OuterVolumeSpecName: "kube-api-access-j9fgq") pod "92a3a8bb-d3df-4489-8cf5-aacdad7516a9" (UID: "92a3a8bb-d3df-4489-8cf5-aacdad7516a9"). InnerVolumeSpecName "kube-api-access-j9fgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.058005 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.058577 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9fgq\" (UniqueName: \"kubernetes.io/projected/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-kube-api-access-j9fgq\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.102377 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92a3a8bb-d3df-4489-8cf5-aacdad7516a9" (UID: "92a3a8bb-d3df-4489-8cf5-aacdad7516a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.162304 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a3a8bb-d3df-4489-8cf5-aacdad7516a9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.268982 4907 generic.go:334] "Generic (PLEG): container finished" podID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" exitCode=0 Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269057 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-598sp" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269025 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880"} Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269135 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-598sp" event={"ID":"92a3a8bb-d3df-4489-8cf5-aacdad7516a9","Type":"ContainerDied","Data":"87a43c517510df3cce3a23c229d4e56247da0404420b2f98fa527a9b52aeadfb"} Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.269220 4907 scope.go:117] "RemoveContainer" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.302494 4907 scope.go:117] "RemoveContainer" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.313745 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.333210 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-598sp"] Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.349001 4907 scope.go:117] "RemoveContainer" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.383995 4907 scope.go:117] "RemoveContainer" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" Mar 13 16:34:26 crc kubenswrapper[4907]: E0313 16:34:26.384879 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880\": container with ID starting with 61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880 not found: ID does not exist" containerID="61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.384987 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880"} err="failed to get container status \"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880\": rpc error: code = NotFound desc = could not find container \"61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880\": container with ID starting with 61eedae48e994ea69752606a6aac87ef010a61cdcce61d46631125a9b841b880 not found: ID does not exist" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.385026 4907 scope.go:117] "RemoveContainer" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" Mar 13 16:34:26 crc kubenswrapper[4907]: E0313 16:34:26.385489 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4\": container with ID starting with 8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4 not found: ID does not exist" containerID="8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.385546 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4"} err="failed to get container status \"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4\": rpc error: code = NotFound desc = could not find container \"8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4\": container with ID starting with 8a3c7e325436c7c09192564adf5ec9964f5d304f4b9cf66a467589dd3de42af4 not found: ID does not exist" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.385580 4907 scope.go:117] "RemoveContainer" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" Mar 13 16:34:26 crc kubenswrapper[4907]: E0313 16:34:26.387311 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a\": container with ID starting with 49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a not found: ID does not exist" containerID="49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.387349 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a"} err="failed to get container status \"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a\": rpc error: code = NotFound desc = could not find container \"49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a\": container with ID starting with 49b82ed2d30c3a4ca01343761e05b4b6d372851890749b3a1174afaa9435ef6a not found: ID does not exist" Mar 13 16:34:26 crc kubenswrapper[4907]: I0313 16:34:26.465352 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.284561 4907 generic.go:334] "Generic (PLEG): container finished" podID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" exitCode=0 Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.284643 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66"} Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.285247 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerStarted","Data":"bce3053aa4e24f9c102b2e439362111ed2eec33f4e6bd2a0aa4f544909c02204"} Mar 13 16:34:27 crc kubenswrapper[4907]: I0313 16:34:27.795090 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" path="/var/lib/kubelet/pods/92a3a8bb-d3df-4489-8cf5-aacdad7516a9/volumes" Mar 13 16:34:28 crc kubenswrapper[4907]: I0313 16:34:28.324389 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerStarted","Data":"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93"} Mar 13 16:34:30 crc kubenswrapper[4907]: I0313 16:34:30.350385 4907 generic.go:334] "Generic (PLEG): container finished" podID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" exitCode=0 Mar 13 16:34:30 crc kubenswrapper[4907]: I0313 16:34:30.350423 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93"} Mar 13 16:34:31 crc kubenswrapper[4907]: I0313 16:34:31.365185 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerStarted","Data":"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb"} Mar 13 16:34:31 crc kubenswrapper[4907]: I0313 16:34:31.402279 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kksrr" podStartSLOduration=2.907058638 podStartE2EDuration="6.402251285s" podCreationTimestamp="2026-03-13 16:34:25 +0000 UTC" firstStartedPulling="2026-03-13 16:34:27.286789392 +0000 UTC m=+8966.186577081" lastFinishedPulling="2026-03-13 16:34:30.781982009 +0000 UTC m=+8969.681769728" observedRunningTime="2026-03-13 16:34:31.387657145 +0000 UTC m=+8970.287444854" watchObservedRunningTime="2026-03-13 16:34:31.402251285 +0000 UTC m=+8970.302039014" Mar 13 16:34:35 crc kubenswrapper[4907]: I0313 16:34:35.895943 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:35 crc kubenswrapper[4907]: I0313 16:34:35.896621 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:36 crc kubenswrapper[4907]: I0313 16:34:36.251535 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:36 crc kubenswrapper[4907]: I0313 16:34:36.491640 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:36 crc kubenswrapper[4907]: I0313 16:34:36.552988 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:38 crc kubenswrapper[4907]: I0313 16:34:38.481093 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kksrr" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" containerID="cri-o://7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" gracePeriod=2 Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.235783 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.402387 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") pod \"b70871f8-ca8b-4909-9d67-b7e720821aea\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.402495 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") pod \"b70871f8-ca8b-4909-9d67-b7e720821aea\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.402595 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") pod \"b70871f8-ca8b-4909-9d67-b7e720821aea\" (UID: \"b70871f8-ca8b-4909-9d67-b7e720821aea\") " Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.403431 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities" (OuterVolumeSpecName: "utilities") pod "b70871f8-ca8b-4909-9d67-b7e720821aea" (UID: "b70871f8-ca8b-4909-9d67-b7e720821aea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.411176 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr" (OuterVolumeSpecName: "kube-api-access-275mr") pod "b70871f8-ca8b-4909-9d67-b7e720821aea" (UID: "b70871f8-ca8b-4909-9d67-b7e720821aea"). InnerVolumeSpecName "kube-api-access-275mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.463848 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b70871f8-ca8b-4909-9d67-b7e720821aea" (UID: "b70871f8-ca8b-4909-9d67-b7e720821aea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494346 4907 generic.go:334] "Generic (PLEG): container finished" podID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" exitCode=0 Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494406 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb"} Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494455 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kksrr" event={"ID":"b70871f8-ca8b-4909-9d67-b7e720821aea","Type":"ContainerDied","Data":"bce3053aa4e24f9c102b2e439362111ed2eec33f4e6bd2a0aa4f544909c02204"} Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.494487 4907 scope.go:117] "RemoveContainer" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.496871 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kksrr" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.505281 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.505323 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b70871f8-ca8b-4909-9d67-b7e720821aea-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.505338 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-275mr\" (UniqueName: \"kubernetes.io/projected/b70871f8-ca8b-4909-9d67-b7e720821aea-kube-api-access-275mr\") on node \"crc\" DevicePath \"\"" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.543118 4907 scope.go:117] "RemoveContainer" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.546998 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.559149 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kksrr"] Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.569493 4907 scope.go:117] "RemoveContainer" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.614703 4907 scope.go:117] "RemoveContainer" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" Mar 13 16:34:39 crc kubenswrapper[4907]: E0313 16:34:39.615150 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb\": container with ID starting with 7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb not found: ID does not exist" containerID="7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615189 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb"} err="failed to get container status \"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb\": rpc error: code = NotFound desc = could not find container \"7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb\": container with ID starting with 7c4d2e7e2ab554003cdeb6b94cee639bd68a9404b66218366440011f09703ffb not found: ID does not exist" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615214 4907 scope.go:117] "RemoveContainer" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" Mar 13 16:34:39 crc kubenswrapper[4907]: E0313 16:34:39.615615 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93\": container with ID starting with 1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93 not found: ID does not exist" containerID="1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615670 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93"} err="failed to get container status \"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93\": rpc error: code = NotFound desc = could not find container \"1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93\": container with ID starting with 1eb85ed4c37b80885e34300324c44dd232a6deab53c7b12db684fc1106d17c93 not found: ID does not exist" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.615687 4907 scope.go:117] "RemoveContainer" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" Mar 13 16:34:39 crc kubenswrapper[4907]: E0313 16:34:39.616105 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66\": container with ID starting with fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66 not found: ID does not exist" containerID="fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.616158 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66"} err="failed to get container status \"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66\": rpc error: code = NotFound desc = could not find container \"fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66\": container with ID starting with fd08cb1e0e69433e63613bf6f882fd195d81484e09fd17ba0e96e23e939edb66 not found: ID does not exist" Mar 13 16:34:39 crc kubenswrapper[4907]: I0313 16:34:39.806274 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" path="/var/lib/kubelet/pods/b70871f8-ca8b-4909-9d67-b7e720821aea/volumes" Mar 13 16:34:43 crc kubenswrapper[4907]: I0313 16:34:43.783213 4907 scope.go:117] "RemoveContainer" containerID="68b2fa8f8bf43cf605cf2bff20ff9f2bc9eb89c3d74246f3836b57e3ae2b5d1c" Mar 13 16:34:48 crc kubenswrapper[4907]: I0313 16:34:48.041410 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:34:48 crc kubenswrapper[4907]: I0313 16:34:48.041832 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:35:18 crc kubenswrapper[4907]: I0313 16:35:18.041871 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:35:18 crc kubenswrapper[4907]: I0313 16:35:18.042312 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.042077 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.042737 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.042799 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.043994 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.044100 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9" gracePeriod=600 Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.405286 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9" exitCode=0 Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.405356 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9"} Mar 13 16:35:48 crc kubenswrapper[4907]: I0313 16:35:48.405707 4907 scope.go:117] "RemoveContainer" containerID="dfb327c310937cae28a687e8c5c0256a2ab50e59a756669f2586112310e1dc69" Mar 13 16:35:49 crc kubenswrapper[4907]: I0313 16:35:49.420714 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047"} Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.209963 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211206 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211225 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211241 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211251 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211277 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211287 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-content" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211297 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211305 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211342 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211352 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: E0313 16:36:00.211377 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211385 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="extract-utilities" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211640 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a3a8bb-d3df-4489-8cf5-aacdad7516a9" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.211671 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b70871f8-ca8b-4909-9d67-b7e720821aea" containerName="registry-server" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.212563 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.227928 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.228294 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.228561 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.244234 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.359028 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"auto-csr-approver-29556996-hhmgk\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.461690 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"auto-csr-approver-29556996-hhmgk\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.483394 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"auto-csr-approver-29556996-hhmgk\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:00 crc kubenswrapper[4907]: I0313 16:36:00.566162 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:01 crc kubenswrapper[4907]: I0313 16:36:01.107303 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:36:01 crc kubenswrapper[4907]: I0313 16:36:01.563569 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" event={"ID":"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5","Type":"ContainerStarted","Data":"8593f1dbfaa7ee1419c6a64af364f636ad809d767e09866e08ca1068367283d8"} Mar 13 16:36:02 crc kubenswrapper[4907]: I0313 16:36:02.578201 4907 generic.go:334] "Generic (PLEG): container finished" podID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerID="fc2a84d7603e9a5913a999c6beb101ceda891932552390c08d77f4d764a6c116" exitCode=0 Mar 13 16:36:02 crc kubenswrapper[4907]: I0313 16:36:02.578277 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" event={"ID":"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5","Type":"ContainerDied","Data":"fc2a84d7603e9a5913a999c6beb101ceda891932552390c08d77f4d764a6c116"} Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.092396 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.256554 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") pod \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\" (UID: \"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5\") " Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.268226 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm" (OuterVolumeSpecName: "kube-api-access-vktpm") pod "d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" (UID: "d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5"). InnerVolumeSpecName "kube-api-access-vktpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.359515 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vktpm\" (UniqueName: \"kubernetes.io/projected/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5-kube-api-access-vktpm\") on node \"crc\" DevicePath \"\"" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.602434 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" event={"ID":"d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5","Type":"ContainerDied","Data":"8593f1dbfaa7ee1419c6a64af364f636ad809d767e09866e08ca1068367283d8"} Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.602473 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556996-hhmgk" Mar 13 16:36:04 crc kubenswrapper[4907]: I0313 16:36:04.602487 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8593f1dbfaa7ee1419c6a64af364f636ad809d767e09866e08ca1068367283d8" Mar 13 16:36:05 crc kubenswrapper[4907]: I0313 16:36:05.167942 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:36:05 crc kubenswrapper[4907]: I0313 16:36:05.183217 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556990-4jrg4"] Mar 13 16:36:05 crc kubenswrapper[4907]: I0313 16:36:05.793961 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a413954-e701-4cec-8868-4fcdaecfabe8" path="/var/lib/kubelet/pods/5a413954-e701-4cec-8868-4fcdaecfabe8/volumes" Mar 13 16:36:08 crc kubenswrapper[4907]: I0313 16:36:08.068713 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 16:36:08 crc kubenswrapper[4907]: I0313 16:36:08.069398 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" containerID="cri-o://be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a" gracePeriod=30 Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.013653 4907 generic.go:334] "Generic (PLEG): container finished" podID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerID="be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a" exitCode=137 Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.014016 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerDied","Data":"be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a"} Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.167768 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.298667 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") pod \"6b679563-ed42-418a-80ef-0b24e225d5ef\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.298785 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") pod \"6b679563-ed42-418a-80ef-0b24e225d5ef\" (UID: \"6b679563-ed42-418a-80ef-0b24e225d5ef\") " Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.307813 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64" (OuterVolumeSpecName: "kube-api-access-rzv64") pod "6b679563-ed42-418a-80ef-0b24e225d5ef" (UID: "6b679563-ed42-418a-80ef-0b24e225d5ef"). InnerVolumeSpecName "kube-api-access-rzv64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.335158 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c" (OuterVolumeSpecName: "mariadb-data") pod "6b679563-ed42-418a-80ef-0b24e225d5ef" (UID: "6b679563-ed42-418a-80ef-0b24e225d5ef"). InnerVolumeSpecName "pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.401824 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") on node \"crc\" " Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.402027 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzv64\" (UniqueName: \"kubernetes.io/projected/6b679563-ed42-418a-80ef-0b24e225d5ef-kube-api-access-rzv64\") on node \"crc\" DevicePath \"\"" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.435219 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.435472 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c") on node "crc" Mar 13 16:36:39 crc kubenswrapper[4907]: I0313 16:36:39.504340 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3b4b3b59-8b2d-44c3-bc88-a79545d23d7c\") on node \"crc\" DevicePath \"\"" Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.031253 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"6b679563-ed42-418a-80ef-0b24e225d5ef","Type":"ContainerDied","Data":"763a1aa9a599cf6e70f4f78d5d2da41195b4bd8f467a89e3d789f85b36605af1"} Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.031334 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.032075 4907 scope.go:117] "RemoveContainer" containerID="be1062c79f4023088d8eb4fb41e0f370f57e80c1813e6265704fdadc605e804a" Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.073424 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.086125 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.823329 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 16:36:40 crc kubenswrapper[4907]: I0313 16:36:40.823662 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" containerID="cri-o://d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" gracePeriod=30 Mar 13 16:36:41 crc kubenswrapper[4907]: I0313 16:36:41.795571 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" path="/var/lib/kubelet/pods/6b679563-ed42-418a-80ef-0b24e225d5ef/volumes" Mar 13 16:36:43 crc kubenswrapper[4907]: I0313 16:36:43.961446 4907 scope.go:117] "RemoveContainer" containerID="5585587b3788c101f53b09ee74a77a65c4d4a4773da4d3e62432619e4599efa1" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.351154 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430298 4907 generic.go:334] "Generic (PLEG): container finished" podID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" exitCode=137 Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430348 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerDied","Data":"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c"} Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430378 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"e0c0f09a-8646-4326-86e8-5bd9123d78ca","Type":"ContainerDied","Data":"afb18a83c45725fa475076eb88bc1f59d47ff74d8c364938ed9892b1ba42b1be"} Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430398 4907 scope.go:117] "RemoveContainer" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.430561 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.459480 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") pod \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.459584 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") pod \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.459689 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") pod \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\" (UID: \"e0c0f09a-8646-4326-86e8-5bd9123d78ca\") " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.461351 4907 scope.go:117] "RemoveContainer" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" Mar 13 16:37:11 crc kubenswrapper[4907]: E0313 16:37:11.462206 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c\": container with ID starting with d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c not found: ID does not exist" containerID="d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.462248 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c"} err="failed to get container status \"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c\": rpc error: code = NotFound desc = could not find container \"d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c\": container with ID starting with d115558c971717b0e8dd16d1b2ccb9b5004c901031eda65c32db4372508a5b7c not found: ID does not exist" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.466435 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2" (OuterVolumeSpecName: "kube-api-access-ctpt2") pod "e0c0f09a-8646-4326-86e8-5bd9123d78ca" (UID: "e0c0f09a-8646-4326-86e8-5bd9123d78ca"). InnerVolumeSpecName "kube-api-access-ctpt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.468796 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "e0c0f09a-8646-4326-86e8-5bd9123d78ca" (UID: "e0c0f09a-8646-4326-86e8-5bd9123d78ca"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.488393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4" (OuterVolumeSpecName: "ovn-data") pod "e0c0f09a-8646-4326-86e8-5bd9123d78ca" (UID: "e0c0f09a-8646-4326-86e8-5bd9123d78ca"). InnerVolumeSpecName "pvc-1a945566-190e-43e6-ae94-58c483d5c2e4". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.562261 4907 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") on node \"crc\" " Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.562292 4907 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/e0c0f09a-8646-4326-86e8-5bd9123d78ca-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.562323 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctpt2\" (UniqueName: \"kubernetes.io/projected/e0c0f09a-8646-4326-86e8-5bd9123d78ca-kube-api-access-ctpt2\") on node \"crc\" DevicePath \"\"" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.588265 4907 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.588829 4907 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1a945566-190e-43e6-ae94-58c483d5c2e4" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4") on node "crc" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.664250 4907 reconciler_common.go:293] "Volume detached for volume \"pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a945566-190e-43e6-ae94-58c483d5c2e4\") on node \"crc\" DevicePath \"\"" Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.777605 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 16:37:11 crc kubenswrapper[4907]: I0313 16:37:11.806389 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Mar 13 16:37:13 crc kubenswrapper[4907]: I0313 16:37:13.811976 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" path="/var/lib/kubelet/pods/e0c0f09a-8646-4326-86e8-5bd9123d78ca/volumes" Mar 13 16:37:48 crc kubenswrapper[4907]: I0313 16:37:48.041439 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:37:48 crc kubenswrapper[4907]: I0313 16:37:48.042195 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.168087 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:38:00 crc kubenswrapper[4907]: E0313 16:38:00.169056 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169071 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: E0313 16:38:00.169107 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169114 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: E0313 16:38:00.169126 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerName="oc" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169132 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerName="oc" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169310 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0c0f09a-8646-4326-86e8-5bd9123d78ca" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169330 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" containerName="oc" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.169347 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b679563-ed42-418a-80ef-0b24e225d5ef" containerName="adoption" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.170101 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.174349 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.174558 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.174713 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.189617 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.370140 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"auto-csr-approver-29556998-pzdj2\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.473068 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"auto-csr-approver-29556998-pzdj2\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.498805 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"auto-csr-approver-29556998-pzdj2\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:00 crc kubenswrapper[4907]: I0313 16:38:00.787738 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:01 crc kubenswrapper[4907]: I0313 16:38:01.110807 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:38:02 crc kubenswrapper[4907]: I0313 16:38:02.055943 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" event={"ID":"2abff695-80c2-4588-9796-44e850ddb65f","Type":"ContainerStarted","Data":"0fce101df987e00b5ae546936da1d9debe45328102484758ec2c10857209676f"} Mar 13 16:38:03 crc kubenswrapper[4907]: I0313 16:38:03.072030 4907 generic.go:334] "Generic (PLEG): container finished" podID="2abff695-80c2-4588-9796-44e850ddb65f" containerID="406f17bcc9c6733e218df39e7a5238412a849d1df878824a54db3db1358e5bff" exitCode=0 Mar 13 16:38:03 crc kubenswrapper[4907]: I0313 16:38:03.072142 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" event={"ID":"2abff695-80c2-4588-9796-44e850ddb65f","Type":"ContainerDied","Data":"406f17bcc9c6733e218df39e7a5238412a849d1df878824a54db3db1358e5bff"} Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.520505 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.693040 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") pod \"2abff695-80c2-4588-9796-44e850ddb65f\" (UID: \"2abff695-80c2-4588-9796-44e850ddb65f\") " Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.698363 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7" (OuterVolumeSpecName: "kube-api-access-zj7j7") pod "2abff695-80c2-4588-9796-44e850ddb65f" (UID: "2abff695-80c2-4588-9796-44e850ddb65f"). InnerVolumeSpecName "kube-api-access-zj7j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:04 crc kubenswrapper[4907]: I0313 16:38:04.796150 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj7j7\" (UniqueName: \"kubernetes.io/projected/2abff695-80c2-4588-9796-44e850ddb65f-kube-api-access-zj7j7\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.105693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" event={"ID":"2abff695-80c2-4588-9796-44e850ddb65f","Type":"ContainerDied","Data":"0fce101df987e00b5ae546936da1d9debe45328102484758ec2c10857209676f"} Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.106082 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fce101df987e00b5ae546936da1d9debe45328102484758ec2c10857209676f" Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.105865 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556998-pzdj2" Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.597374 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.607185 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556992-qj2mm"] Mar 13 16:38:05 crc kubenswrapper[4907]: I0313 16:38:05.793962 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d2dfd3f-0bc6-48f9-85f7-100ef8f23476" path="/var/lib/kubelet/pods/5d2dfd3f-0bc6-48f9-85f7-100ef8f23476/volumes" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.638298 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:38:08 crc kubenswrapper[4907]: E0313 16:38:08.639257 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2abff695-80c2-4588-9796-44e850ddb65f" containerName="oc" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.639270 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2abff695-80c2-4588-9796-44e850ddb65f" containerName="oc" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.639464 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2abff695-80c2-4588-9796-44e850ddb65f" containerName="oc" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.640625 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.642450 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-xndl6"/"default-dockercfg-m9xng" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.642955 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xndl6"/"kube-root-ca.crt" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.643136 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-xndl6"/"openshift-service-ca.crt" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.647225 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.786728 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.786898 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.889166 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.889937 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.891944 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.915050 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"must-gather-8f2lx\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:08 crc kubenswrapper[4907]: I0313 16:38:08.962122 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:38:09 crc kubenswrapper[4907]: I0313 16:38:09.479498 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:38:10 crc kubenswrapper[4907]: I0313 16:38:10.159100 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerStarted","Data":"269636f3f131040884ec72a96a5ea7fa20184677b56a7a261e911c974a24da8b"} Mar 13 16:38:17 crc kubenswrapper[4907]: I0313 16:38:17.272151 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerStarted","Data":"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f"} Mar 13 16:38:17 crc kubenswrapper[4907]: I0313 16:38:17.272603 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerStarted","Data":"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19"} Mar 13 16:38:17 crc kubenswrapper[4907]: I0313 16:38:17.289402 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xndl6/must-gather-8f2lx" podStartSLOduration=2.20949555 podStartE2EDuration="9.289386532s" podCreationTimestamp="2026-03-13 16:38:08 +0000 UTC" firstStartedPulling="2026-03-13 16:38:09.475788773 +0000 UTC m=+9188.375576452" lastFinishedPulling="2026-03-13 16:38:16.555679755 +0000 UTC m=+9195.455467434" observedRunningTime="2026-03-13 16:38:17.288638001 +0000 UTC m=+9196.188425690" watchObservedRunningTime="2026-03-13 16:38:17.289386532 +0000 UTC m=+9196.189174221" Mar 13 16:38:18 crc kubenswrapper[4907]: I0313 16:38:18.041487 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:38:18 crc kubenswrapper[4907]: I0313 16:38:18.041737 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:38:20 crc kubenswrapper[4907]: I0313 16:38:20.941940 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xndl6/crc-debug-qt756"] Mar 13 16:38:20 crc kubenswrapper[4907]: I0313 16:38:20.944199 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.069500 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.069785 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.171617 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.171725 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.172093 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.189444 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"crc-debug-qt756\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.263065 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:21 crc kubenswrapper[4907]: I0313 16:38:21.310076 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-qt756" event={"ID":"5e5575c9-0bdc-472b-9384-fbd1ce5cb480","Type":"ContainerStarted","Data":"f3e01960410a5bad3f4d4d1123a16e5c7cc67a5d538faaf418827fff78354c63"} Mar 13 16:38:34 crc kubenswrapper[4907]: I0313 16:38:34.457779 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-qt756" event={"ID":"5e5575c9-0bdc-472b-9384-fbd1ce5cb480","Type":"ContainerStarted","Data":"b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90"} Mar 13 16:38:34 crc kubenswrapper[4907]: I0313 16:38:34.476091 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-xndl6/crc-debug-qt756" podStartSLOduration=2.283363906 podStartE2EDuration="14.476070234s" podCreationTimestamp="2026-03-13 16:38:20 +0000 UTC" firstStartedPulling="2026-03-13 16:38:21.299905065 +0000 UTC m=+9200.199692754" lastFinishedPulling="2026-03-13 16:38:33.492611393 +0000 UTC m=+9212.392399082" observedRunningTime="2026-03-13 16:38:34.471182577 +0000 UTC m=+9213.370970266" watchObservedRunningTime="2026-03-13 16:38:34.476070234 +0000 UTC m=+9213.375857923" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.546546 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.549416 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.561683 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.602551 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.602606 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.602631 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.704579 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.704855 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.705019 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.705361 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.705372 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.723484 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"redhat-marketplace-vz6rb\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:39 crc kubenswrapper[4907]: I0313 16:38:39.867481 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.107160 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:42 crc kubenswrapper[4907]: W0313 16:38:42.139329 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b31d6ac_276a_4947_a468_e99a784e1ff3.slice/crio-6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89 WatchSource:0}: Error finding container 6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89: Status 404 returned error can't find the container with id 6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89 Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.537514 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" exitCode=0 Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.537614 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124"} Mar 13 16:38:42 crc kubenswrapper[4907]: I0313 16:38:42.537912 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerStarted","Data":"6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89"} Mar 13 16:38:43 crc kubenswrapper[4907]: I0313 16:38:43.551301 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerStarted","Data":"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95"} Mar 13 16:38:44 crc kubenswrapper[4907]: I0313 16:38:44.133986 4907 scope.go:117] "RemoveContainer" containerID="a1001aa55a6abdb3dd8474ef6137b7a29548724c2eff0d7fa57396148bbc916a" Mar 13 16:38:44 crc kubenswrapper[4907]: I0313 16:38:44.562473 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" exitCode=0 Mar 13 16:38:44 crc kubenswrapper[4907]: I0313 16:38:44.562786 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95"} Mar 13 16:38:45 crc kubenswrapper[4907]: I0313 16:38:45.575601 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerStarted","Data":"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd"} Mar 13 16:38:45 crc kubenswrapper[4907]: I0313 16:38:45.604123 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vz6rb" podStartSLOduration=3.988551902 podStartE2EDuration="6.604100368s" podCreationTimestamp="2026-03-13 16:38:39 +0000 UTC" firstStartedPulling="2026-03-13 16:38:42.539860052 +0000 UTC m=+9221.439647741" lastFinishedPulling="2026-03-13 16:38:45.155408518 +0000 UTC m=+9224.055196207" observedRunningTime="2026-03-13 16:38:45.597362618 +0000 UTC m=+9224.497150307" watchObservedRunningTime="2026-03-13 16:38:45.604100368 +0000 UTC m=+9224.503888047" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.041453 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.043078 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.043223 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.044236 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.044413 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" gracePeriod=600 Mar 13 16:38:48 crc kubenswrapper[4907]: E0313 16:38:48.179659 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.606463 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" exitCode=0 Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.606515 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047"} Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.606834 4907 scope.go:117] "RemoveContainer" containerID="f9156a2a4d423d000cb09f67e16772f675217bb2e483ba02adf57b642ffdffc9" Mar 13 16:38:48 crc kubenswrapper[4907]: I0313 16:38:48.607749 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:38:48 crc kubenswrapper[4907]: E0313 16:38:48.608108 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:38:49 crc kubenswrapper[4907]: I0313 16:38:49.868031 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:49 crc kubenswrapper[4907]: I0313 16:38:49.868519 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:49 crc kubenswrapper[4907]: I0313 16:38:49.934542 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:50 crc kubenswrapper[4907]: I0313 16:38:50.677865 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:50 crc kubenswrapper[4907]: I0313 16:38:50.733274 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:51 crc kubenswrapper[4907]: I0313 16:38:51.636413 4907 generic.go:334] "Generic (PLEG): container finished" podID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerID="b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90" exitCode=0 Mar 13 16:38:51 crc kubenswrapper[4907]: I0313 16:38:51.636492 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-qt756" event={"ID":"5e5575c9-0bdc-472b-9384-fbd1ce5cb480","Type":"ContainerDied","Data":"b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90"} Mar 13 16:38:52 crc kubenswrapper[4907]: I0313 16:38:52.648181 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vz6rb" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" containerID="cri-o://a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" gracePeriod=2 Mar 13 16:38:52 crc kubenswrapper[4907]: I0313 16:38:52.944337 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:52 crc kubenswrapper[4907]: I0313 16:38:52.995152 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-qt756"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.004550 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-qt756"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017296 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") pod \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017358 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") pod \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\" (UID: \"5e5575c9-0bdc-472b-9384-fbd1ce5cb480\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017633 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host" (OuterVolumeSpecName: "host") pod "5e5575c9-0bdc-472b-9384-fbd1ce5cb480" (UID: "5e5575c9-0bdc-472b-9384-fbd1ce5cb480"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.017833 4907 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-host\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.032134 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv" (OuterVolumeSpecName: "kube-api-access-4vjlv") pod "5e5575c9-0bdc-472b-9384-fbd1ce5cb480" (UID: "5e5575c9-0bdc-472b-9384-fbd1ce5cb480"). InnerVolumeSpecName "kube-api-access-4vjlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.120568 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vjlv\" (UniqueName: \"kubernetes.io/projected/5e5575c9-0bdc-472b-9384-fbd1ce5cb480-kube-api-access-4vjlv\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.227733 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.324249 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") pod \"2b31d6ac-276a-4947-a468-e99a784e1ff3\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.324564 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") pod \"2b31d6ac-276a-4947-a468-e99a784e1ff3\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.324599 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") pod \"2b31d6ac-276a-4947-a468-e99a784e1ff3\" (UID: \"2b31d6ac-276a-4947-a468-e99a784e1ff3\") " Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.325393 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities" (OuterVolumeSpecName: "utilities") pod "2b31d6ac-276a-4947-a468-e99a784e1ff3" (UID: "2b31d6ac-276a-4947-a468-e99a784e1ff3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.329011 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl" (OuterVolumeSpecName: "kube-api-access-9qbcl") pod "2b31d6ac-276a-4947-a468-e99a784e1ff3" (UID: "2b31d6ac-276a-4947-a468-e99a784e1ff3"). InnerVolumeSpecName "kube-api-access-9qbcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.362420 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b31d6ac-276a-4947-a468-e99a784e1ff3" (UID: "2b31d6ac-276a-4947-a468-e99a784e1ff3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.426960 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.427011 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qbcl\" (UniqueName: \"kubernetes.io/projected/2b31d6ac-276a-4947-a468-e99a784e1ff3-kube-api-access-9qbcl\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.427027 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b31d6ac-276a-4947-a468-e99a784e1ff3-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.657513 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3e01960410a5bad3f4d4d1123a16e5c7cc67a5d538faaf418827fff78354c63" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.657576 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-qt756" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662345 4907 generic.go:334] "Generic (PLEG): container finished" podID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" exitCode=0 Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662391 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd"} Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662421 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6rb" event={"ID":"2b31d6ac-276a-4947-a468-e99a784e1ff3","Type":"ContainerDied","Data":"6aff65ef2fefe4fdcba78cafb69943816320eedcf9d24e99d5c0c1143b301c89"} Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662438 4907 scope.go:117] "RemoveContainer" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.662572 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6rb" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.720675 4907 scope.go:117] "RemoveContainer" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.754498 4907 scope.go:117] "RemoveContainer" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.769659 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.779775 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6rb"] Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.795388 4907 scope.go:117] "RemoveContainer" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" Mar 13 16:38:53 crc kubenswrapper[4907]: E0313 16:38:53.802230 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd\": container with ID starting with a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd not found: ID does not exist" containerID="a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.802267 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd"} err="failed to get container status \"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd\": rpc error: code = NotFound desc = could not find container \"a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd\": container with ID starting with a54681c8803fa507d1b76e516386bee6ee700712c69fdb45137bd6c2f0e3e8cd not found: ID does not exist" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.802345 4907 scope.go:117] "RemoveContainer" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" Mar 13 16:38:53 crc kubenswrapper[4907]: E0313 16:38:53.803000 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95\": container with ID starting with fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95 not found: ID does not exist" containerID="fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.803041 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95"} err="failed to get container status \"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95\": rpc error: code = NotFound desc = could not find container \"fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95\": container with ID starting with fd46e261e524d075378035d76f6c3b9852c9358bf16927c6ec958aaaf314bb95 not found: ID does not exist" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.803069 4907 scope.go:117] "RemoveContainer" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" Mar 13 16:38:53 crc kubenswrapper[4907]: E0313 16:38:53.803554 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124\": container with ID starting with 201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124 not found: ID does not exist" containerID="201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.803584 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124"} err="failed to get container status \"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124\": rpc error: code = NotFound desc = could not find container \"201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124\": container with ID starting with 201fc9ca5006a0bcb74f754932e0f1754a8156f25796f4eac96197c357dad124 not found: ID does not exist" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.811695 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" path="/var/lib/kubelet/pods/2b31d6ac-276a-4947-a468-e99a784e1ff3/volumes" Mar 13 16:38:53 crc kubenswrapper[4907]: I0313 16:38:53.812439 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" path="/var/lib/kubelet/pods/5e5575c9-0bdc-472b-9384-fbd1ce5cb480/volumes" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.215950 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-xndl6/crc-debug-gr5ml"] Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216523 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216540 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216561 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-content" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216567 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-content" Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216593 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-utilities" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216599 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="extract-utilities" Mar 13 16:38:54 crc kubenswrapper[4907]: E0313 16:38:54.216618 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerName="container-00" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216624 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerName="container-00" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216800 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5575c9-0bdc-472b-9384-fbd1ce5cb480" containerName="container-00" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.216820 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b31d6ac-276a-4947-a468-e99a784e1ff3" containerName="registry-server" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.217516 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.373821 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.373925 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.476402 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.476503 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.476659 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.502594 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"crc-debug-gr5ml\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.536106 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:54 crc kubenswrapper[4907]: I0313 16:38:54.673382 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" event={"ID":"09517507-85bc-48d4-9fe2-f4a9e862322a","Type":"ContainerStarted","Data":"d004f3ada10244b8dd65e6d6ddad786d1dfdd84dd59f991b7610b3493b6f4b3e"} Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.687824 4907 generic.go:334] "Generic (PLEG): container finished" podID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerID="97e8dbb17dfb954f08537b16b9fb37c2f747c102c1339b2be558bedc609e5946" exitCode=1 Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.688137 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" event={"ID":"09517507-85bc-48d4-9fe2-f4a9e862322a","Type":"ContainerDied","Data":"97e8dbb17dfb954f08537b16b9fb37c2f747c102c1339b2be558bedc609e5946"} Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.732595 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-gr5ml"] Mar 13 16:38:55 crc kubenswrapper[4907]: I0313 16:38:55.741743 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xndl6/crc-debug-gr5ml"] Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.002694 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.029668 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") pod \"09517507-85bc-48d4-9fe2-f4a9e862322a\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.029756 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") pod \"09517507-85bc-48d4-9fe2-f4a9e862322a\" (UID: \"09517507-85bc-48d4-9fe2-f4a9e862322a\") " Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.029917 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host" (OuterVolumeSpecName: "host") pod "09517507-85bc-48d4-9fe2-f4a9e862322a" (UID: "09517507-85bc-48d4-9fe2-f4a9e862322a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.030279 4907 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/09517507-85bc-48d4-9fe2-f4a9e862322a-host\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.035160 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2" (OuterVolumeSpecName: "kube-api-access-v5sf2") pod "09517507-85bc-48d4-9fe2-f4a9e862322a" (UID: "09517507-85bc-48d4-9fe2-f4a9e862322a"). InnerVolumeSpecName "kube-api-access-v5sf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.133343 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5sf2\" (UniqueName: \"kubernetes.io/projected/09517507-85bc-48d4-9fe2-f4a9e862322a-kube-api-access-v5sf2\") on node \"crc\" DevicePath \"\"" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.711655 4907 scope.go:117] "RemoveContainer" containerID="97e8dbb17dfb954f08537b16b9fb37c2f747c102c1339b2be558bedc609e5946" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.711660 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/crc-debug-gr5ml" Mar 13 16:38:57 crc kubenswrapper[4907]: I0313 16:38:57.793961 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" path="/var/lib/kubelet/pods/09517507-85bc-48d4-9fe2-f4a9e862322a/volumes" Mar 13 16:38:59 crc kubenswrapper[4907]: I0313 16:38:59.785913 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:38:59 crc kubenswrapper[4907]: E0313 16:38:59.786799 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:11 crc kubenswrapper[4907]: I0313 16:39:11.790052 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:11 crc kubenswrapper[4907]: E0313 16:39:11.790942 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:22 crc kubenswrapper[4907]: I0313 16:39:22.782553 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:22 crc kubenswrapper[4907]: E0313 16:39:22.783963 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:37 crc kubenswrapper[4907]: I0313 16:39:37.783189 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:37 crc kubenswrapper[4907]: E0313 16:39:37.784137 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:39:52 crc kubenswrapper[4907]: I0313 16:39:52.783620 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:39:52 crc kubenswrapper[4907]: E0313 16:39:52.784987 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.155043 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:40:00 crc kubenswrapper[4907]: E0313 16:40:00.161760 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerName="container-00" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.161798 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerName="container-00" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.164644 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="09517507-85bc-48d4-9fe2-f4a9e862322a" containerName="container-00" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.166059 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.170410 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.170770 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.181930 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.235375 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.375275 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"auto-csr-approver-29557000-b8z24\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.477614 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"auto-csr-approver-29557000-b8z24\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.499398 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"auto-csr-approver-29557000-b8z24\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:00 crc kubenswrapper[4907]: I0313 16:40:00.511189 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:01 crc kubenswrapper[4907]: I0313 16:40:01.029157 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:40:01 crc kubenswrapper[4907]: I0313 16:40:01.043482 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:40:01 crc kubenswrapper[4907]: I0313 16:40:01.625216 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerStarted","Data":"3975471bc6fb2efe82da83ec7e799b1b9e617df03d5750fb3e8165d894b63258"} Mar 13 16:40:02 crc kubenswrapper[4907]: I0313 16:40:02.638523 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerStarted","Data":"7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e"} Mar 13 16:40:02 crc kubenswrapper[4907]: I0313 16:40:02.665772 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557000-b8z24" podStartSLOduration=1.5020899349999999 podStartE2EDuration="2.665746056s" podCreationTimestamp="2026-03-13 16:40:00 +0000 UTC" firstStartedPulling="2026-03-13 16:40:01.043089806 +0000 UTC m=+9299.942877525" lastFinishedPulling="2026-03-13 16:40:02.206745957 +0000 UTC m=+9301.106533646" observedRunningTime="2026-03-13 16:40:02.654380796 +0000 UTC m=+9301.554168525" watchObservedRunningTime="2026-03-13 16:40:02.665746056 +0000 UTC m=+9301.565533785" Mar 13 16:40:03 crc kubenswrapper[4907]: I0313 16:40:03.650590 4907 generic.go:334] "Generic (PLEG): container finished" podID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerID="7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e" exitCode=0 Mar 13 16:40:03 crc kubenswrapper[4907]: I0313 16:40:03.650637 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerDied","Data":"7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e"} Mar 13 16:40:03 crc kubenswrapper[4907]: I0313 16:40:03.782964 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:03 crc kubenswrapper[4907]: E0313 16:40:03.783388 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.054271 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.201020 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") pod \"ff06fab3-a77c-49ef-8c13-88908b7acf21\" (UID: \"ff06fab3-a77c-49ef-8c13-88908b7acf21\") " Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.206999 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q" (OuterVolumeSpecName: "kube-api-access-9cb9q") pod "ff06fab3-a77c-49ef-8c13-88908b7acf21" (UID: "ff06fab3-a77c-49ef-8c13-88908b7acf21"). InnerVolumeSpecName "kube-api-access-9cb9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.304095 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cb9q\" (UniqueName: \"kubernetes.io/projected/ff06fab3-a77c-49ef-8c13-88908b7acf21-kube-api-access-9cb9q\") on node \"crc\" DevicePath \"\"" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.672861 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557000-b8z24" event={"ID":"ff06fab3-a77c-49ef-8c13-88908b7acf21","Type":"ContainerDied","Data":"3975471bc6fb2efe82da83ec7e799b1b9e617df03d5750fb3e8165d894b63258"} Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.673342 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3975471bc6fb2efe82da83ec7e799b1b9e617df03d5750fb3e8165d894b63258" Mar 13 16:40:05 crc kubenswrapper[4907]: I0313 16:40:05.672951 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557000-b8z24" Mar 13 16:40:06 crc kubenswrapper[4907]: I0313 16:40:06.129233 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:40:06 crc kubenswrapper[4907]: I0313 16:40:06.138653 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556994-452h2"] Mar 13 16:40:07 crc kubenswrapper[4907]: I0313 16:40:07.807609 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="262afeb8-16de-47e7-a64b-5b4e938aa1f4" path="/var/lib/kubelet/pods/262afeb8-16de-47e7-a64b-5b4e938aa1f4/volumes" Mar 13 16:40:15 crc kubenswrapper[4907]: I0313 16:40:15.782623 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:15 crc kubenswrapper[4907]: E0313 16:40:15.783397 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:30 crc kubenswrapper[4907]: I0313 16:40:30.783304 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:30 crc kubenswrapper[4907]: E0313 16:40:30.784238 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:42 crc kubenswrapper[4907]: I0313 16:40:42.782911 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:42 crc kubenswrapper[4907]: E0313 16:40:42.783752 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:40:44 crc kubenswrapper[4907]: I0313 16:40:44.289348 4907 scope.go:117] "RemoveContainer" containerID="237bb05fcc8f490c8ced9263406f8715e053a689a6c929dacb3684e153ea9c27" Mar 13 16:40:55 crc kubenswrapper[4907]: I0313 16:40:55.782829 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:40:55 crc kubenswrapper[4907]: E0313 16:40:55.783739 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:09 crc kubenswrapper[4907]: I0313 16:41:09.783391 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:09 crc kubenswrapper[4907]: E0313 16:41:09.784394 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:20 crc kubenswrapper[4907]: I0313 16:41:20.783652 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:20 crc kubenswrapper[4907]: E0313 16:41:20.784950 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:34 crc kubenswrapper[4907]: I0313 16:41:34.791798 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:34 crc kubenswrapper[4907]: E0313 16:41:34.793390 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:41:47 crc kubenswrapper[4907]: I0313 16:41:47.782781 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:41:47 crc kubenswrapper[4907]: E0313 16:41:47.783542 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.156333 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:42:00 crc kubenswrapper[4907]: E0313 16:42:00.157648 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerName="oc" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.157665 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerName="oc" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.157979 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" containerName="oc" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.158961 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.172272 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.192348 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.192483 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.192556 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.225358 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"auto-csr-approver-29557002-tdngs\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.328204 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"auto-csr-approver-29557002-tdngs\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.355156 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"auto-csr-approver-29557002-tdngs\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:00 crc kubenswrapper[4907]: I0313 16:42:00.510376 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:01 crc kubenswrapper[4907]: I0313 16:42:01.684669 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:42:01 crc kubenswrapper[4907]: I0313 16:42:01.790277 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:01 crc kubenswrapper[4907]: E0313 16:42:01.790520 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:01 crc kubenswrapper[4907]: I0313 16:42:01.852324 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557002-tdngs" event={"ID":"dce83493-bbc5-49ac-a420-229eac93005d","Type":"ContainerStarted","Data":"cc8a87d050d7d4532f011d62079dba28815de9c2ad3f8a91b6523667a402eab2"} Mar 13 16:42:04 crc kubenswrapper[4907]: I0313 16:42:04.897038 4907 generic.go:334] "Generic (PLEG): container finished" podID="dce83493-bbc5-49ac-a420-229eac93005d" containerID="3ca53b3677f234a39cf1706a0419909431a4370ca20c691d14a98c1cf5bf1d3b" exitCode=0 Mar 13 16:42:04 crc kubenswrapper[4907]: I0313 16:42:04.897257 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557002-tdngs" event={"ID":"dce83493-bbc5-49ac-a420-229eac93005d","Type":"ContainerDied","Data":"3ca53b3677f234a39cf1706a0419909431a4370ca20c691d14a98c1cf5bf1d3b"} Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.356931 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.486836 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") pod \"dce83493-bbc5-49ac-a420-229eac93005d\" (UID: \"dce83493-bbc5-49ac-a420-229eac93005d\") " Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.495594 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l" (OuterVolumeSpecName: "kube-api-access-fbw8l") pod "dce83493-bbc5-49ac-a420-229eac93005d" (UID: "dce83493-bbc5-49ac-a420-229eac93005d"). InnerVolumeSpecName "kube-api-access-fbw8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.589808 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbw8l\" (UniqueName: \"kubernetes.io/projected/dce83493-bbc5-49ac-a420-229eac93005d-kube-api-access-fbw8l\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.923519 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557002-tdngs" event={"ID":"dce83493-bbc5-49ac-a420-229eac93005d","Type":"ContainerDied","Data":"cc8a87d050d7d4532f011d62079dba28815de9c2ad3f8a91b6523667a402eab2"} Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.923581 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc8a87d050d7d4532f011d62079dba28815de9c2ad3f8a91b6523667a402eab2" Mar 13 16:42:06 crc kubenswrapper[4907]: I0313 16:42:06.923658 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557002-tdngs" Mar 13 16:42:07 crc kubenswrapper[4907]: I0313 16:42:07.449207 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:42:07 crc kubenswrapper[4907]: I0313 16:42:07.459730 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556996-hhmgk"] Mar 13 16:42:07 crc kubenswrapper[4907]: I0313 16:42:07.798839 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5" path="/var/lib/kubelet/pods/d0ba1c68-f3a1-4c88-bc6b-db7c438dd2f5/volumes" Mar 13 16:42:15 crc kubenswrapper[4907]: I0313 16:42:15.783840 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:15 crc kubenswrapper[4907]: E0313 16:42:15.785195 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.277590 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:20 crc kubenswrapper[4907]: E0313 16:42:20.278686 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dce83493-bbc5-49ac-a420-229eac93005d" containerName="oc" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.278702 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="dce83493-bbc5-49ac-a420-229eac93005d" containerName="oc" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.279018 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="dce83493-bbc5-49ac-a420-229eac93005d" containerName="oc" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.280924 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.287997 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.368301 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.368379 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.368473 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.470428 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.470777 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.470834 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.471032 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.471313 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.491875 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"redhat-operators-s2qr6\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:20 crc kubenswrapper[4907]: I0313 16:42:20.634187 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:21 crc kubenswrapper[4907]: I0313 16:42:21.132206 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:22 crc kubenswrapper[4907]: I0313 16:42:22.103067 4907 generic.go:334] "Generic (PLEG): container finished" podID="20c838c2-685d-4fa7-bde6-a88be618698f" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" exitCode=0 Mar 13 16:42:22 crc kubenswrapper[4907]: I0313 16:42:22.103371 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e"} Mar 13 16:42:22 crc kubenswrapper[4907]: I0313 16:42:22.103401 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerStarted","Data":"2a340d1af9caff0371f5bfd63f2a8323b61e8d31999ba5f5f637876a00501768"} Mar 13 16:42:24 crc kubenswrapper[4907]: I0313 16:42:24.129314 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerStarted","Data":"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26"} Mar 13 16:42:28 crc kubenswrapper[4907]: I0313 16:42:28.181186 4907 generic.go:334] "Generic (PLEG): container finished" podID="20c838c2-685d-4fa7-bde6-a88be618698f" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" exitCode=0 Mar 13 16:42:28 crc kubenswrapper[4907]: I0313 16:42:28.181261 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26"} Mar 13 16:42:28 crc kubenswrapper[4907]: I0313 16:42:28.783805 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:28 crc kubenswrapper[4907]: E0313 16:42:28.784988 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.209521 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerStarted","Data":"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5"} Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.249364 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s2qr6" podStartSLOduration=3.191336437 podStartE2EDuration="10.249343034s" podCreationTimestamp="2026-03-13 16:42:20 +0000 UTC" firstStartedPulling="2026-03-13 16:42:22.107003149 +0000 UTC m=+9441.006790838" lastFinishedPulling="2026-03-13 16:42:29.165009716 +0000 UTC m=+9448.064797435" observedRunningTime="2026-03-13 16:42:30.23210169 +0000 UTC m=+9449.131889389" watchObservedRunningTime="2026-03-13 16:42:30.249343034 +0000 UTC m=+9449.149130733" Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.634577 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:30 crc kubenswrapper[4907]: I0313 16:42:30.634701 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:31 crc kubenswrapper[4907]: I0313 16:42:31.695617 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s2qr6" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" probeResult="failure" output=< Mar 13 16:42:31 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:42:31 crc kubenswrapper[4907]: > Mar 13 16:42:40 crc kubenswrapper[4907]: I0313 16:42:40.698596 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:40 crc kubenswrapper[4907]: I0313 16:42:40.762854 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:40 crc kubenswrapper[4907]: I0313 16:42:40.952280 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:42 crc kubenswrapper[4907]: I0313 16:42:42.524203 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s2qr6" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" containerID="cri-o://619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" gracePeriod=2 Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.111916 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.223468 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") pod \"20c838c2-685d-4fa7-bde6-a88be618698f\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.223568 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") pod \"20c838c2-685d-4fa7-bde6-a88be618698f\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.224450 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") pod \"20c838c2-685d-4fa7-bde6-a88be618698f\" (UID: \"20c838c2-685d-4fa7-bde6-a88be618698f\") " Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.224992 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities" (OuterVolumeSpecName: "utilities") pod "20c838c2-685d-4fa7-bde6-a88be618698f" (UID: "20c838c2-685d-4fa7-bde6-a88be618698f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.231133 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq" (OuterVolumeSpecName: "kube-api-access-qhmlq") pod "20c838c2-685d-4fa7-bde6-a88be618698f" (UID: "20c838c2-685d-4fa7-bde6-a88be618698f"). InnerVolumeSpecName "kube-api-access-qhmlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.326950 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.326996 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhmlq\" (UniqueName: \"kubernetes.io/projected/20c838c2-685d-4fa7-bde6-a88be618698f-kube-api-access-qhmlq\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.361681 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20c838c2-685d-4fa7-bde6-a88be618698f" (UID: "20c838c2-685d-4fa7-bde6-a88be618698f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.428388 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20c838c2-685d-4fa7-bde6-a88be618698f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537782 4907 generic.go:334] "Generic (PLEG): container finished" podID="20c838c2-685d-4fa7-bde6-a88be618698f" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" exitCode=0 Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537828 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5"} Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537855 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s2qr6" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537875 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s2qr6" event={"ID":"20c838c2-685d-4fa7-bde6-a88be618698f","Type":"ContainerDied","Data":"2a340d1af9caff0371f5bfd63f2a8323b61e8d31999ba5f5f637876a00501768"} Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.537992 4907 scope.go:117] "RemoveContainer" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.572602 4907 scope.go:117] "RemoveContainer" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.595062 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.606631 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s2qr6"] Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.612172 4907 scope.go:117] "RemoveContainer" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.654920 4907 scope.go:117] "RemoveContainer" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.655742 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5\": container with ID starting with 619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5 not found: ID does not exist" containerID="619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.655825 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5"} err="failed to get container status \"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5\": rpc error: code = NotFound desc = could not find container \"619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5\": container with ID starting with 619dc8758167775cf4458460d2573597c7ee1b199f5844f41d3d53c7ecb99ab5 not found: ID does not exist" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.655868 4907 scope.go:117] "RemoveContainer" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.656384 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26\": container with ID starting with 167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26 not found: ID does not exist" containerID="167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.656451 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26"} err="failed to get container status \"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26\": rpc error: code = NotFound desc = could not find container \"167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26\": container with ID starting with 167fc2eea8c311e949891f542a0aa9abeb88120496908b41c8378abfe2846f26 not found: ID does not exist" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.656492 4907 scope.go:117] "RemoveContainer" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.657622 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e\": container with ID starting with 0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e not found: ID does not exist" containerID="0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.657661 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e"} err="failed to get container status \"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e\": rpc error: code = NotFound desc = could not find container \"0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e\": container with ID starting with 0c270278e59534e523470376925881213319928159d9e96183cc72cf4f76ee2e not found: ID does not exist" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.783396 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:43 crc kubenswrapper[4907]: E0313 16:42:43.784181 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:42:43 crc kubenswrapper[4907]: I0313 16:42:43.800576 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" path="/var/lib/kubelet/pods/20c838c2-685d-4fa7-bde6-a88be618698f/volumes" Mar 13 16:42:44 crc kubenswrapper[4907]: I0313 16:42:44.427829 4907 scope.go:117] "RemoveContainer" containerID="fc2a84d7603e9a5913a999c6beb101ceda891932552390c08d77f4d764a6c116" Mar 13 16:42:54 crc kubenswrapper[4907]: I0313 16:42:54.783108 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:42:54 crc kubenswrapper[4907]: E0313 16:42:54.783928 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:06 crc kubenswrapper[4907]: I0313 16:43:06.783084 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:06 crc kubenswrapper[4907]: E0313 16:43:06.784035 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:17 crc kubenswrapper[4907]: I0313 16:43:17.783517 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:17 crc kubenswrapper[4907]: E0313 16:43:17.784585 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:28 crc kubenswrapper[4907]: I0313 16:43:28.782781 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:28 crc kubenswrapper[4907]: E0313 16:43:28.783587 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:40 crc kubenswrapper[4907]: I0313 16:43:40.783168 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:40 crc kubenswrapper[4907]: E0313 16:43:40.784149 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:43:51 crc kubenswrapper[4907]: I0313 16:43:51.788675 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:43:52 crc kubenswrapper[4907]: I0313 16:43:52.231562 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9"} Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.148279 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:44:00 crc kubenswrapper[4907]: E0313 16:44:00.149671 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-utilities" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.149695 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-utilities" Mar 13 16:44:00 crc kubenswrapper[4907]: E0313 16:44:00.149729 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-content" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.149741 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="extract-content" Mar 13 16:44:00 crc kubenswrapper[4907]: E0313 16:44:00.149795 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.149804 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.150120 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c838c2-685d-4fa7-bde6-a88be618698f" containerName="registry-server" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.151157 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.154669 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.154840 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.155023 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.164415 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.279054 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"auto-csr-approver-29557004-v4vjf\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.381385 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"auto-csr-approver-29557004-v4vjf\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.418511 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"auto-csr-approver-29557004-v4vjf\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:00 crc kubenswrapper[4907]: I0313 16:44:00.476029 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:01 crc kubenswrapper[4907]: I0313 16:44:01.018024 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:44:01 crc kubenswrapper[4907]: W0313 16:44:01.020806 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod724a5965_1261_4da5_a833_40c8af08fde6.slice/crio-c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db WatchSource:0}: Error finding container c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db: Status 404 returned error can't find the container with id c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db Mar 13 16:44:01 crc kubenswrapper[4907]: I0313 16:44:01.336826 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerStarted","Data":"c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db"} Mar 13 16:44:03 crc kubenswrapper[4907]: I0313 16:44:03.371706 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerStarted","Data":"656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845"} Mar 13 16:44:03 crc kubenswrapper[4907]: I0313 16:44:03.408680 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" podStartSLOduration=2.5342579819999997 podStartE2EDuration="3.408648985s" podCreationTimestamp="2026-03-13 16:44:00 +0000 UTC" firstStartedPulling="2026-03-13 16:44:01.023102398 +0000 UTC m=+9539.922890117" lastFinishedPulling="2026-03-13 16:44:01.897493431 +0000 UTC m=+9540.797281120" observedRunningTime="2026-03-13 16:44:03.392752988 +0000 UTC m=+9542.292540707" watchObservedRunningTime="2026-03-13 16:44:03.408648985 +0000 UTC m=+9542.308436684" Mar 13 16:44:04 crc kubenswrapper[4907]: I0313 16:44:04.392108 4907 generic.go:334] "Generic (PLEG): container finished" podID="724a5965-1261-4da5-a833-40c8af08fde6" containerID="656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845" exitCode=0 Mar 13 16:44:04 crc kubenswrapper[4907]: I0313 16:44:04.392228 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerDied","Data":"656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845"} Mar 13 16:44:05 crc kubenswrapper[4907]: I0313 16:44:05.882018 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.039976 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") pod \"724a5965-1261-4da5-a833-40c8af08fde6\" (UID: \"724a5965-1261-4da5-a833-40c8af08fde6\") " Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.053325 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2" (OuterVolumeSpecName: "kube-api-access-bdlp2") pod "724a5965-1261-4da5-a833-40c8af08fde6" (UID: "724a5965-1261-4da5-a833-40c8af08fde6"). InnerVolumeSpecName "kube-api-access-bdlp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.143157 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdlp2\" (UniqueName: \"kubernetes.io/projected/724a5965-1261-4da5-a833-40c8af08fde6-kube-api-access-bdlp2\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.427332 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" event={"ID":"724a5965-1261-4da5-a833-40c8af08fde6","Type":"ContainerDied","Data":"c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db"} Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.427380 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c033b71ea5e962cba3a37e561238afdcae9f063023456a9b5197facd2b95a8db" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.427443 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557004-v4vjf" Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.487250 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:44:06 crc kubenswrapper[4907]: I0313 16:44:06.495688 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556998-pzdj2"] Mar 13 16:44:07 crc kubenswrapper[4907]: I0313 16:44:07.808797 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2abff695-80c2-4588-9796-44e850ddb65f" path="/var/lib/kubelet/pods/2abff695-80c2-4588-9796-44e850ddb65f/volumes" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.283857 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:33 crc kubenswrapper[4907]: E0313 16:44:33.285757 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="724a5965-1261-4da5-a833-40c8af08fde6" containerName="oc" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.285776 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="724a5965-1261-4da5-a833-40c8af08fde6" containerName="oc" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.286145 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="724a5965-1261-4da5-a833-40c8af08fde6" containerName="oc" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.288093 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.302206 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.342630 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.343137 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.343229 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.444821 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445081 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445191 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445307 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.445445 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.470672 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"certified-operators-b9df4\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:33 crc kubenswrapper[4907]: I0313 16:44:33.634229 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.136859 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.821873 4907 generic.go:334] "Generic (PLEG): container finished" podID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" exitCode=0 Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.821936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4"} Mar 13 16:44:34 crc kubenswrapper[4907]: I0313 16:44:34.822002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerStarted","Data":"1e5271fea52afdf222fc307f42b571fe15871546b42ce2ccc995b3c6f4ae18ff"} Mar 13 16:44:35 crc kubenswrapper[4907]: I0313 16:44:35.835395 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerStarted","Data":"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b"} Mar 13 16:44:37 crc kubenswrapper[4907]: I0313 16:44:37.864810 4907 generic.go:334] "Generic (PLEG): container finished" podID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" exitCode=0 Mar 13 16:44:37 crc kubenswrapper[4907]: I0313 16:44:37.864945 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b"} Mar 13 16:44:38 crc kubenswrapper[4907]: I0313 16:44:38.877893 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerStarted","Data":"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825"} Mar 13 16:44:38 crc kubenswrapper[4907]: I0313 16:44:38.901470 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b9df4" podStartSLOduration=2.400423634 podStartE2EDuration="5.901456055s" podCreationTimestamp="2026-03-13 16:44:33 +0000 UTC" firstStartedPulling="2026-03-13 16:44:34.829745512 +0000 UTC m=+9573.729533221" lastFinishedPulling="2026-03-13 16:44:38.330777943 +0000 UTC m=+9577.230565642" observedRunningTime="2026-03-13 16:44:38.898317287 +0000 UTC m=+9577.798104986" watchObservedRunningTime="2026-03-13 16:44:38.901456055 +0000 UTC m=+9577.801243744" Mar 13 16:44:43 crc kubenswrapper[4907]: I0313 16:44:43.748925 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:43 crc kubenswrapper[4907]: I0313 16:44:43.749602 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.369664 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.426000 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.571126 4907 scope.go:117] "RemoveContainer" containerID="b1d7df69f877a557d28aa9c5dd1537366e05d49104f113b20536b953f8c2ec90" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.616320 4907 scope.go:117] "RemoveContainer" containerID="406f17bcc9c6733e218df39e7a5238412a849d1df878824a54db3db1358e5bff" Mar 13 16:44:44 crc kubenswrapper[4907]: I0313 16:44:44.632363 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:45 crc kubenswrapper[4907]: I0313 16:44:45.955405 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b9df4" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" containerID="cri-o://9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" gracePeriod=2 Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.489285 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.652845 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") pod \"b1428c7e-574e-4fa0-a364-14a6c85d413a\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.653035 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") pod \"b1428c7e-574e-4fa0-a364-14a6c85d413a\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.653103 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") pod \"b1428c7e-574e-4fa0-a364-14a6c85d413a\" (UID: \"b1428c7e-574e-4fa0-a364-14a6c85d413a\") " Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.653958 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities" (OuterVolumeSpecName: "utilities") pod "b1428c7e-574e-4fa0-a364-14a6c85d413a" (UID: "b1428c7e-574e-4fa0-a364-14a6c85d413a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.659895 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv" (OuterVolumeSpecName: "kube-api-access-2m8vv") pod "b1428c7e-574e-4fa0-a364-14a6c85d413a" (UID: "b1428c7e-574e-4fa0-a364-14a6c85d413a"). InnerVolumeSpecName "kube-api-access-2m8vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.755692 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m8vv\" (UniqueName: \"kubernetes.io/projected/b1428c7e-574e-4fa0-a364-14a6c85d413a-kube-api-access-2m8vv\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.755744 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971137 4907 generic.go:334] "Generic (PLEG): container finished" podID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" exitCode=0 Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971554 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825"} Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9df4" event={"ID":"b1428c7e-574e-4fa0-a364-14a6c85d413a","Type":"ContainerDied","Data":"1e5271fea52afdf222fc307f42b571fe15871546b42ce2ccc995b3c6f4ae18ff"} Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971617 4907 scope.go:117] "RemoveContainer" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" Mar 13 16:44:46 crc kubenswrapper[4907]: I0313 16:44:46.971776 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9df4" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.002209 4907 scope.go:117] "RemoveContainer" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.035765 4907 scope.go:117] "RemoveContainer" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.070116 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1428c7e-574e-4fa0-a364-14a6c85d413a" (UID: "b1428c7e-574e-4fa0-a364-14a6c85d413a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.096674 4907 scope.go:117] "RemoveContainer" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" Mar 13 16:44:47 crc kubenswrapper[4907]: E0313 16:44:47.097202 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825\": container with ID starting with 9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825 not found: ID does not exist" containerID="9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.097266 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825"} err="failed to get container status \"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825\": rpc error: code = NotFound desc = could not find container \"9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825\": container with ID starting with 9083630a661a76583aad7aa2abb1771ff314d367addcdb10ff8aca627760d825 not found: ID does not exist" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.097299 4907 scope.go:117] "RemoveContainer" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" Mar 13 16:44:47 crc kubenswrapper[4907]: E0313 16:44:47.098300 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b\": container with ID starting with c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b not found: ID does not exist" containerID="c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.098346 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b"} err="failed to get container status \"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b\": rpc error: code = NotFound desc = could not find container \"c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b\": container with ID starting with c70795cf0ff18f40beec2e966464d6e8d8e8e805f466ccd0e9155b99894fdc8b not found: ID does not exist" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.098373 4907 scope.go:117] "RemoveContainer" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" Mar 13 16:44:47 crc kubenswrapper[4907]: E0313 16:44:47.098870 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4\": container with ID starting with 406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4 not found: ID does not exist" containerID="406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.098928 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4"} err="failed to get container status \"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4\": rpc error: code = NotFound desc = could not find container \"406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4\": container with ID starting with 406b700ec04569af0a43aa0f00c007b21e68df9b9a2e35767629619ce91ea8f4 not found: ID does not exist" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.179634 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1428c7e-574e-4fa0-a364-14a6c85d413a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.308256 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.320299 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b9df4"] Mar 13 16:44:47 crc kubenswrapper[4907]: I0313 16:44:47.815721 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" path="/var/lib/kubelet/pods/b1428c7e-574e-4fa0-a364-14a6c85d413a/volumes" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.158209 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r"] Mar 13 16:45:00 crc kubenswrapper[4907]: E0313 16:45:00.159146 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159161 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" Mar 13 16:45:00 crc kubenswrapper[4907]: E0313 16:45:00.159176 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-content" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159185 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-content" Mar 13 16:45:00 crc kubenswrapper[4907]: E0313 16:45:00.159212 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-utilities" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159221 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="extract-utilities" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.159497 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1428c7e-574e-4fa0-a364-14a6c85d413a" containerName="registry-server" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.160471 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.163846 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.166630 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.171977 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r"] Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.308726 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.309435 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.309512 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.411483 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.412059 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.412305 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.412857 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.421006 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.431816 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"collect-profiles-29557005-4k95r\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:00 crc kubenswrapper[4907]: I0313 16:45:00.481064 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:01 crc kubenswrapper[4907]: I0313 16:45:01.001127 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r"] Mar 13 16:45:01 crc kubenswrapper[4907]: I0313 16:45:01.145767 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" event={"ID":"5a94884c-c438-4536-baa4-fefad79add1a","Type":"ContainerStarted","Data":"1dea5c9209ab95881c1b9078c066ab6d0957cdb5d7b2ac2325232d2ef8b8cc41"} Mar 13 16:45:02 crc kubenswrapper[4907]: I0313 16:45:02.156537 4907 generic.go:334] "Generic (PLEG): container finished" podID="5a94884c-c438-4536-baa4-fefad79add1a" containerID="5f6d7d4723c79a2c85bc8ce346672905a252464fdb7c28aa0ce2166188a31bd1" exitCode=0 Mar 13 16:45:02 crc kubenswrapper[4907]: I0313 16:45:02.156654 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" event={"ID":"5a94884c-c438-4536-baa4-fefad79add1a","Type":"ContainerDied","Data":"5f6d7d4723c79a2c85bc8ce346672905a252464fdb7c28aa0ce2166188a31bd1"} Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.629646 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.705090 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") pod \"5a94884c-c438-4536-baa4-fefad79add1a\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.705274 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") pod \"5a94884c-c438-4536-baa4-fefad79add1a\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.705967 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume" (OuterVolumeSpecName: "config-volume") pod "5a94884c-c438-4536-baa4-fefad79add1a" (UID: "5a94884c-c438-4536-baa4-fefad79add1a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.706052 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") pod \"5a94884c-c438-4536-baa4-fefad79add1a\" (UID: \"5a94884c-c438-4536-baa4-fefad79add1a\") " Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.706517 4907 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5a94884c-c438-4536-baa4-fefad79add1a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.712856 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq" (OuterVolumeSpecName: "kube-api-access-z5gnq") pod "5a94884c-c438-4536-baa4-fefad79add1a" (UID: "5a94884c-c438-4536-baa4-fefad79add1a"). InnerVolumeSpecName "kube-api-access-z5gnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.713413 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5a94884c-c438-4536-baa4-fefad79add1a" (UID: "5a94884c-c438-4536-baa4-fefad79add1a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.815506 4907 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5a94884c-c438-4536-baa4-fefad79add1a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:03 crc kubenswrapper[4907]: I0313 16:45:03.815539 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5gnq\" (UniqueName: \"kubernetes.io/projected/5a94884c-c438-4536-baa4-fefad79add1a-kube-api-access-z5gnq\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.183299 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" event={"ID":"5a94884c-c438-4536-baa4-fefad79add1a","Type":"ContainerDied","Data":"1dea5c9209ab95881c1b9078c066ab6d0957cdb5d7b2ac2325232d2ef8b8cc41"} Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.183564 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dea5c9209ab95881c1b9078c066ab6d0957cdb5d7b2ac2325232d2ef8b8cc41" Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.183380 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29557005-4k95r" Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.741059 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:45:04 crc kubenswrapper[4907]: I0313 16:45:04.753856 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556960-mkbxd"] Mar 13 16:45:05 crc kubenswrapper[4907]: I0313 16:45:05.805385 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1001d4c-6984-4a3c-8d1f-167a6a48ea8a" path="/var/lib/kubelet/pods/e1001d4c-6984-4a3c-8d1f-167a6a48ea8a/volumes" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.357439 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:31 crc kubenswrapper[4907]: E0313 16:45:31.359935 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a94884c-c438-4536-baa4-fefad79add1a" containerName="collect-profiles" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.359971 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a94884c-c438-4536-baa4-fefad79add1a" containerName="collect-profiles" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.360470 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a94884c-c438-4536-baa4-fefad79add1a" containerName="collect-profiles" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.364068 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.396721 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.440188 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.440269 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.440585 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.543833 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.543933 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.543974 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.544846 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.544843 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.572681 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"community-operators-wx4sr\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:31 crc kubenswrapper[4907]: I0313 16:45:31.725072 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:32 crc kubenswrapper[4907]: I0313 16:45:32.319245 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:32 crc kubenswrapper[4907]: W0313 16:45:32.324104 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod864d645e_16d7_4d3b_82a1_d763ee19600d.slice/crio-a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89 WatchSource:0}: Error finding container a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89: Status 404 returned error can't find the container with id a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89 Mar 13 16:45:32 crc kubenswrapper[4907]: I0313 16:45:32.577449 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0"} Mar 13 16:45:32 crc kubenswrapper[4907]: I0313 16:45:32.577524 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89"} Mar 13 16:45:33 crc kubenswrapper[4907]: I0313 16:45:33.592454 4907 generic.go:334] "Generic (PLEG): container finished" podID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" exitCode=0 Mar 13 16:45:33 crc kubenswrapper[4907]: I0313 16:45:33.592578 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0"} Mar 13 16:45:33 crc kubenswrapper[4907]: I0313 16:45:33.596752 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:45:35 crc kubenswrapper[4907]: I0313 16:45:35.632300 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1"} Mar 13 16:45:37 crc kubenswrapper[4907]: I0313 16:45:37.656512 4907 generic.go:334] "Generic (PLEG): container finished" podID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" exitCode=0 Mar 13 16:45:37 crc kubenswrapper[4907]: I0313 16:45:37.656560 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1"} Mar 13 16:45:38 crc kubenswrapper[4907]: I0313 16:45:38.671640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerStarted","Data":"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b"} Mar 13 16:45:38 crc kubenswrapper[4907]: I0313 16:45:38.708116 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wx4sr" podStartSLOduration=3.244016488 podStartE2EDuration="7.708090076s" podCreationTimestamp="2026-03-13 16:45:31 +0000 UTC" firstStartedPulling="2026-03-13 16:45:33.596477629 +0000 UTC m=+9632.496265328" lastFinishedPulling="2026-03-13 16:45:38.060551207 +0000 UTC m=+9636.960338916" observedRunningTime="2026-03-13 16:45:38.694124943 +0000 UTC m=+9637.593912652" watchObservedRunningTime="2026-03-13 16:45:38.708090076 +0000 UTC m=+9637.607877785" Mar 13 16:45:41 crc kubenswrapper[4907]: I0313 16:45:41.832967 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:41 crc kubenswrapper[4907]: I0313 16:45:41.833580 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:42 crc kubenswrapper[4907]: I0313 16:45:42.866268 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wx4sr" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" probeResult="failure" output=< Mar 13 16:45:42 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:45:42 crc kubenswrapper[4907]: > Mar 13 16:45:44 crc kubenswrapper[4907]: I0313 16:45:44.772583 4907 scope.go:117] "RemoveContainer" containerID="0c0e8b2fe415ff36180c0f0cbc79e145510c4e8b607c452be50ff3ed2394d856" Mar 13 16:45:51 crc kubenswrapper[4907]: I0313 16:45:51.804479 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:51 crc kubenswrapper[4907]: I0313 16:45:51.887268 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:52 crc kubenswrapper[4907]: I0313 16:45:52.397101 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:52 crc kubenswrapper[4907]: I0313 16:45:52.836983 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wx4sr" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" containerID="cri-o://4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" gracePeriod=2 Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.445020 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.554070 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") pod \"864d645e-16d7-4d3b-82a1-d763ee19600d\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.554405 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") pod \"864d645e-16d7-4d3b-82a1-d763ee19600d\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.554848 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") pod \"864d645e-16d7-4d3b-82a1-d763ee19600d\" (UID: \"864d645e-16d7-4d3b-82a1-d763ee19600d\") " Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.555414 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities" (OuterVolumeSpecName: "utilities") pod "864d645e-16d7-4d3b-82a1-d763ee19600d" (UID: "864d645e-16d7-4d3b-82a1-d763ee19600d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.555800 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.561188 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7" (OuterVolumeSpecName: "kube-api-access-hf9w7") pod "864d645e-16d7-4d3b-82a1-d763ee19600d" (UID: "864d645e-16d7-4d3b-82a1-d763ee19600d"). InnerVolumeSpecName "kube-api-access-hf9w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.613196 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "864d645e-16d7-4d3b-82a1-d763ee19600d" (UID: "864d645e-16d7-4d3b-82a1-d763ee19600d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.657598 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf9w7\" (UniqueName: \"kubernetes.io/projected/864d645e-16d7-4d3b-82a1-d763ee19600d-kube-api-access-hf9w7\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.657629 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/864d645e-16d7-4d3b-82a1-d763ee19600d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856723 4907 generic.go:334] "Generic (PLEG): container finished" podID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" exitCode=0 Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856764 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b"} Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856792 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wx4sr" event={"ID":"864d645e-16d7-4d3b-82a1-d763ee19600d","Type":"ContainerDied","Data":"a307370b5b3ddfe67ecb5180a2f6aea9904a507dc0c7d87ae8a5138047252a89"} Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856812 4907 scope.go:117] "RemoveContainer" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.856918 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wx4sr" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.895271 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.899072 4907 scope.go:117] "RemoveContainer" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.904083 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wx4sr"] Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.937572 4907 scope.go:117] "RemoveContainer" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.978492 4907 scope.go:117] "RemoveContainer" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" Mar 13 16:45:53 crc kubenswrapper[4907]: E0313 16:45:53.979107 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b\": container with ID starting with 4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b not found: ID does not exist" containerID="4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979158 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b"} err="failed to get container status \"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b\": rpc error: code = NotFound desc = could not find container \"4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b\": container with ID starting with 4a312a3cf18fab398fbf6c5eaa5295284d64d484a558188758c899a3134be59b not found: ID does not exist" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979194 4907 scope.go:117] "RemoveContainer" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" Mar 13 16:45:53 crc kubenswrapper[4907]: E0313 16:45:53.979587 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1\": container with ID starting with 38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1 not found: ID does not exist" containerID="38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979627 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1"} err="failed to get container status \"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1\": rpc error: code = NotFound desc = could not find container \"38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1\": container with ID starting with 38540a9b7efd6bb2a5b35b5d21a8bacf08f2ab0da9a4f992cf84f289f305c6c1 not found: ID does not exist" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.979653 4907 scope.go:117] "RemoveContainer" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" Mar 13 16:45:53 crc kubenswrapper[4907]: E0313 16:45:53.980001 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0\": container with ID starting with ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0 not found: ID does not exist" containerID="ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0" Mar 13 16:45:53 crc kubenswrapper[4907]: I0313 16:45:53.980037 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0"} err="failed to get container status \"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0\": rpc error: code = NotFound desc = could not find container \"ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0\": container with ID starting with ecb06530e4c6cf8e5865dec7c587d6387a6ebfa7dea5833ab58b5054e66656b0 not found: ID does not exist" Mar 13 16:45:55 crc kubenswrapper[4907]: I0313 16:45:55.799367 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" path="/var/lib/kubelet/pods/864d645e-16d7-4d3b-82a1-d763ee19600d/volumes" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.198596 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:46:00 crc kubenswrapper[4907]: E0313 16:46:00.199692 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-content" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.199707 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-content" Mar 13 16:46:00 crc kubenswrapper[4907]: E0313 16:46:00.199731 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-utilities" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.199739 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="extract-utilities" Mar 13 16:46:00 crc kubenswrapper[4907]: E0313 16:46:00.199748 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.199758 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.200046 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="864d645e-16d7-4d3b-82a1-d763ee19600d" containerName="registry-server" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.200917 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.201001 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.206711 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.208394 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.208779 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.339523 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"auto-csr-approver-29557006-ck2rj\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.442692 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"auto-csr-approver-29557006-ck2rj\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.473120 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"auto-csr-approver-29557006-ck2rj\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:00 crc kubenswrapper[4907]: I0313 16:46:00.528163 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:01 crc kubenswrapper[4907]: I0313 16:46:01.027159 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:46:01 crc kubenswrapper[4907]: I0313 16:46:01.973296 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" event={"ID":"0690b5f5-ba57-4fa5-baf4-86a729917817","Type":"ContainerStarted","Data":"d7f0cf0bc5d44cbbf4ef528b69bdcaff5527f17372aaf7b2a0cab117b0fe4f68"} Mar 13 16:46:02 crc kubenswrapper[4907]: I0313 16:46:02.989794 4907 generic.go:334] "Generic (PLEG): container finished" podID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerID="29e87ba308a2bea52f09c17c0673a7492d85691ad7ee7034046f4e32cf2bff88" exitCode=0 Mar 13 16:46:02 crc kubenswrapper[4907]: I0313 16:46:02.989927 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" event={"ID":"0690b5f5-ba57-4fa5-baf4-86a729917817","Type":"ContainerDied","Data":"29e87ba308a2bea52f09c17c0673a7492d85691ad7ee7034046f4e32cf2bff88"} Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.519377 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.675628 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") pod \"0690b5f5-ba57-4fa5-baf4-86a729917817\" (UID: \"0690b5f5-ba57-4fa5-baf4-86a729917817\") " Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.682109 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb" (OuterVolumeSpecName: "kube-api-access-r9swb") pod "0690b5f5-ba57-4fa5-baf4-86a729917817" (UID: "0690b5f5-ba57-4fa5-baf4-86a729917817"). InnerVolumeSpecName "kube-api-access-r9swb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:46:04 crc kubenswrapper[4907]: I0313 16:46:04.778257 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9swb\" (UniqueName: \"kubernetes.io/projected/0690b5f5-ba57-4fa5-baf4-86a729917817-kube-api-access-r9swb\") on node \"crc\" DevicePath \"\"" Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.021418 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" event={"ID":"0690b5f5-ba57-4fa5-baf4-86a729917817","Type":"ContainerDied","Data":"d7f0cf0bc5d44cbbf4ef528b69bdcaff5527f17372aaf7b2a0cab117b0fe4f68"} Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.021822 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7f0cf0bc5d44cbbf4ef528b69bdcaff5527f17372aaf7b2a0cab117b0fe4f68" Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.021518 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557006-ck2rj" Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.612142 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.622967 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557000-b8z24"] Mar 13 16:46:05 crc kubenswrapper[4907]: I0313 16:46:05.806320 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff06fab3-a77c-49ef-8c13-88908b7acf21" path="/var/lib/kubelet/pods/ff06fab3-a77c-49ef-8c13-88908b7acf21/volumes" Mar 13 16:46:18 crc kubenswrapper[4907]: I0313 16:46:18.041708 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:46:18 crc kubenswrapper[4907]: I0313 16:46:18.042450 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:46:44 crc kubenswrapper[4907]: I0313 16:46:44.852857 4907 scope.go:117] "RemoveContainer" containerID="7edb971c3f915cfb8576dc1860f2e0063b3420e5f584dbd096705d5ce436c71e" Mar 13 16:46:48 crc kubenswrapper[4907]: I0313 16:46:48.042559 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:46:48 crc kubenswrapper[4907]: I0313 16:46:48.044835 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.041919 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.042542 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.042610 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.043748 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.043851 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9" gracePeriod=600 Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.976914 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9" exitCode=0 Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.977002 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9"} Mar 13 16:47:18 crc kubenswrapper[4907]: I0313 16:47:18.977270 4907 scope.go:117] "RemoveContainer" containerID="7e4d0690f4ac607ea8a268bfa88b5ae6d9a95f8b6013de7747fdffb429f28047" Mar 13 16:47:19 crc kubenswrapper[4907]: I0313 16:47:19.987776 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc"} Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.140214 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:48:00 crc kubenswrapper[4907]: E0313 16:48:00.141179 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerName="oc" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.141194 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerName="oc" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.141403 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" containerName="oc" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.142230 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.144057 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.144544 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.144698 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.150852 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.220011 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"auto-csr-approver-29557008-tqdp7\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.322409 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"auto-csr-approver-29557008-tqdp7\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.343400 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"auto-csr-approver-29557008-tqdp7\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.459915 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:00 crc kubenswrapper[4907]: I0313 16:48:00.923496 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:48:01 crc kubenswrapper[4907]: I0313 16:48:01.608582 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerStarted","Data":"7b6ed12b3c6ca6c521826924c61118611556bd303b2b9c0e550e05e23de1591e"} Mar 13 16:48:03 crc kubenswrapper[4907]: I0313 16:48:03.627493 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerStarted","Data":"feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31"} Mar 13 16:48:03 crc kubenswrapper[4907]: I0313 16:48:03.647486 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" podStartSLOduration=1.512070255 podStartE2EDuration="3.647467128s" podCreationTimestamp="2026-03-13 16:48:00 +0000 UTC" firstStartedPulling="2026-03-13 16:48:00.942505104 +0000 UTC m=+9779.842292833" lastFinishedPulling="2026-03-13 16:48:03.077902017 +0000 UTC m=+9781.977689706" observedRunningTime="2026-03-13 16:48:03.642917443 +0000 UTC m=+9782.542705142" watchObservedRunningTime="2026-03-13 16:48:03.647467128 +0000 UTC m=+9782.547254817" Mar 13 16:48:04 crc kubenswrapper[4907]: I0313 16:48:04.642266 4907 generic.go:334] "Generic (PLEG): container finished" podID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerID="feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31" exitCode=0 Mar 13 16:48:04 crc kubenswrapper[4907]: I0313 16:48:04.643640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerDied","Data":"feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31"} Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.132446 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.253944 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") pod \"168ac198-4031-4bf7-bcf6-eb190b28dce6\" (UID: \"168ac198-4031-4bf7-bcf6-eb190b28dce6\") " Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.264185 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4" (OuterVolumeSpecName: "kube-api-access-sjbk4") pod "168ac198-4031-4bf7-bcf6-eb190b28dce6" (UID: "168ac198-4031-4bf7-bcf6-eb190b28dce6"). InnerVolumeSpecName "kube-api-access-sjbk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.360203 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjbk4\" (UniqueName: \"kubernetes.io/projected/168ac198-4031-4bf7-bcf6-eb190b28dce6-kube-api-access-sjbk4\") on node \"crc\" DevicePath \"\"" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.662429 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" event={"ID":"168ac198-4031-4bf7-bcf6-eb190b28dce6","Type":"ContainerDied","Data":"7b6ed12b3c6ca6c521826924c61118611556bd303b2b9c0e550e05e23de1591e"} Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.662704 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b6ed12b3c6ca6c521826924c61118611556bd303b2b9c0e550e05e23de1591e" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.662727 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557008-tqdp7" Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.734063 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:48:06 crc kubenswrapper[4907]: I0313 16:48:06.745715 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557002-tdngs"] Mar 13 16:48:07 crc kubenswrapper[4907]: I0313 16:48:07.800927 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dce83493-bbc5-49ac-a420-229eac93005d" path="/var/lib/kubelet/pods/dce83493-bbc5-49ac-a420-229eac93005d/volumes" Mar 13 16:48:19 crc kubenswrapper[4907]: I0313 16:48:19.088227 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/init-config-reloader/0.log" Mar 13 16:48:19 crc kubenswrapper[4907]: I0313 16:48:19.966844 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/init-config-reloader/0.log" Mar 13 16:48:19 crc kubenswrapper[4907]: I0313 16:48:19.989975 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/alertmanager/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.031995 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_d29f6de0-3229-477c-86a6-550f2d465592/config-reloader/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.178257 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-api/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.239183 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-evaluator/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.269061 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-listener/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.357724 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_8ce53a96-2566-4d5c-b4cd-bbfd2516975a/aodh-notifier/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.454220 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5564857dbd-vn7xm_d83fca2b-ed84-4516-8510-6e0ff4b2cfef/barbican-api/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.517841 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5564857dbd-vn7xm_d83fca2b-ed84-4516-8510-6e0ff4b2cfef/barbican-api-log/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.680628 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68d9445c58-bjlh2_87b4d8b9-57da-4c61-9dd9-d746fc62db84/barbican-keystone-listener/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.701511 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68d9445c58-bjlh2_87b4d8b9-57da-4c61-9dd9-d746fc62db84/barbican-keystone-listener-log/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.939094 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-868b698d7f-wbtl4_3599175f-0bbd-4645-ab11-07e26d7fb98d/barbican-worker-log/0.log" Mar 13 16:48:20 crc kubenswrapper[4907]: I0313 16:48:20.962366 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-868b698d7f-wbtl4_3599175f-0bbd-4645-ab11-07e26d7fb98d/barbican-worker/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.063182 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-tb9sw_d7346ccb-6154-42a9-9097-bad966681d81/bootstrap-openstack-openstack-cell1/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.313269 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/ceilometer-notification-agent/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.325402 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/ceilometer-central-agent/0.log" Mar 13 16:48:21 crc kubenswrapper[4907]: I0313 16:48:21.450923 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/proxy-httpd/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.225928 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_59fb2cd9-5414-4356-9438-d33f90ad084c/sg-core/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.417289 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-tcvvp_ef037fca-09ee-48ed-b2f7-fc65aea9155a/ceph-client-openstack-openstack-cell1/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.922733 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8c82a596-df2b-4e69-abe9-5f4662a3c6da/cinder-api-log/0.log" Mar 13 16:48:22 crc kubenswrapper[4907]: I0313 16:48:22.970339 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8c82a596-df2b-4e69-abe9-5f4662a3c6da/cinder-api/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.107854 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d92d5136-2169-44cc-9900-c3335a4b69eb/cinder-backup/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.292553 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_d92d5136-2169-44cc-9900-c3335a4b69eb/probe/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.356338 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3a999cb1-5df0-43ae-946f-eae3f2c53848/cinder-scheduler/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.364433 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3a999cb1-5df0-43ae-946f-eae3f2c53848/probe/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.628120 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_697762c8-b121-410a-a5db-21916ca1791f/probe/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.633705 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_697762c8-b121-410a-a5db-21916ca1791f/cinder-volume/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.727061 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-wrvtc_bfe0612b-2d05-4625-b14e-8b9ea02a8c06/configure-network-openstack-openstack-cell1/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.919070 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-s9llf_173f7de2-ef0c-4750-a8b3-23e078ab9bde/configure-os-openstack-openstack-cell1/0.log" Mar 13 16:48:23 crc kubenswrapper[4907]: I0313 16:48:23.999520 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8644fdb9df-gm62m_44fcf282-3688-4a6f-b390-6b321e89b158/init/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.208798 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8644fdb9df-gm62m_44fcf282-3688-4a6f-b390-6b321e89b158/dnsmasq-dns/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.240536 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8644fdb9df-gm62m_44fcf282-3688-4a6f-b390-6b321e89b158/init/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.270579 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-84zm8_441cfe9f-03ed-4a4f-b3f6-58ff77c45cc9/download-cache-openstack-openstack-cell1/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.488115 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_95435ffc-5322-4bd8-9558-421f953096eb/glance-httpd/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.524992 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_70b550c7-5c28-416d-83c3-1ba63ecd5abd/glance-httpd/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.548216 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_70b550c7-5c28-416d-83c3-1ba63ecd5abd/glance-log/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.581487 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_95435ffc-5322-4bd8-9558-421f953096eb/glance-log/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.828681 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7fd69496d-xglm5_a76ba2c0-c077-44ad-a2ed-2ee4f1a0052a/heat-api/0.log" Mar 13 16:48:24 crc kubenswrapper[4907]: I0313 16:48:24.981743 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-7dbbf4d5b8-846ft_1d7cc945-4d68-41c8-9ea0-359af3279cc8/heat-cfnapi/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.059347 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-86cbbf5dd8-25c77_3dd72cd1-7342-40aa-8c3c-9d14375b3930/heat-engine/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.260856 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d79dff969-ssj4t_5315ef54-fd95-400c-a7ef-a356c0f867bc/horizon/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.264224 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6d79dff969-ssj4t_5315ef54-fd95-400c-a7ef-a356c0f867bc/horizon-log/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.321443 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-55pht_c32eadac-12ec-49b8-930e-bc6af99631e8/install-certs-openstack-openstack-cell1/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.464249 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-vw9h5_cc30c8d7-5f7e-47c8-b89f-6a638b886769/install-os-openstack-openstack-cell1/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.714740 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-58b5f55f68-nlrbn_121ed53e-99ad-41a2-b8e6-56e4f56520c6/keystone-api/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.726947 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29556961-gshbz_e548a313-fc14-4ce9-b268-bf37c79eed83/keystone-cron/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.771611 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_53c58ef1-fd3c-4d26-9b00-5ce0bb0cc592/kube-state-metrics/0.log" Mar 13 16:48:25 crc kubenswrapper[4907]: I0313 16:48:25.979148 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-bdwbh_82c7c5cb-4f6b-4d38-a009-263025faa45b/libvirt-openstack-openstack-cell1/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.135336 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f2518181-dd4d-4d35-9377-061bbd91c811/manila-api/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.207101 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_f2518181-dd4d-4d35-9377-061bbd91c811/manila-api-log/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.459386 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_2e179ecf-0fc5-4675-89a6-287d3398dc2d/probe/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.533541 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_93911045-021f-489b-9817-f0a24066c906/manila-share/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.585730 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_2e179ecf-0fc5-4675-89a6-287d3398dc2d/manila-scheduler/0.log" Mar 13 16:48:26 crc kubenswrapper[4907]: I0313 16:48:26.646992 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_93911045-021f-489b-9817-f0a24066c906/probe/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.035095 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-564986d4bf-zhh4c_3ce663fe-b525-44ea-a6d4-33f6b2366f46/neutron-httpd/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.037271 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-564986d4bf-zhh4c_3ce663fe-b525-44ea-a6d4-33f6b2366f46/neutron-api/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.150345 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-jmbms_b5764492-ad4a-4e43-8a98-99ea7cfd9248/neutron-dhcp-openstack-openstack-cell1/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.397641 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-pwqj6_3dc18cfe-d63a-4960-8c7d-10913cc78896/neutron-metadata-openstack-openstack-cell1/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.568057 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-sqxhp_05d384d7-9635-4d68-8eaf-48fea8bed0b8/neutron-sriov-openstack-openstack-cell1/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.805115 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c11f341d-9a76-4e7e-b1e7-8d63c30bac32/nova-api-api/0.log" Mar 13 16:48:27 crc kubenswrapper[4907]: I0313 16:48:27.956119 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c11f341d-9a76-4e7e-b1e7-8d63c30bac32/nova-api-log/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.096873 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_fd1f3838-01fe-4e37-9859-8aa28048e81c/nova-cell0-conductor-conductor/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.257674 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0a6bc64d-334f-46cd-aea9-486ab56d4cf9/nova-cell1-conductor-conductor/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.548121 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_317b6fd4-e642-4858-bc3d-2cce128ea0e7/nova-cell1-novncproxy-novncproxy/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.590245 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell22qwc_de68abba-24e7-49b8-9e1f-adfa05d9d3dd/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.731672 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-2swxs_563c85ca-38c8-4853-a278-2a46fd0347a3/nova-cell1-openstack-openstack-cell1/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.864290 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_13364227-6148-4b0d-ba7d-082578dfe1d5/nova-metadata-metadata/0.log" Mar 13 16:48:28 crc kubenswrapper[4907]: I0313 16:48:28.950618 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_13364227-6148-4b0d-ba7d-082578dfe1d5/nova-metadata-log/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.110338 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_54123701-97fb-4f4f-baed-4c2183bad203/nova-scheduler-scheduler/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.697929 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/init/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.936501 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/init/0.log" Mar 13 16:48:29 crc kubenswrapper[4907]: I0313 16:48:29.942104 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/octavia-api-provider-agent/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.213480 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-6db876fcbd-4wtkj_1c868ec4-5e49-4be1-a1eb-c4748753aecb/octavia-api/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.369252 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hbbw4_57610eac-b9eb-4628-bd99-b3ebb35b2772/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.629500 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hbbw4_57610eac-b9eb-4628-bd99-b3ebb35b2772/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.641344 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-p69c5_ab5868a8-f083-4417-b7c0-f01dad538927/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.712224 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-hbbw4_57610eac-b9eb-4628-bd99-b3ebb35b2772/octavia-healthmanager/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.884290 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-p69c5_ab5868a8-f083-4417-b7c0-f01dad538927/init/0.log" Mar 13 16:48:30 crc kubenswrapper[4907]: I0313 16:48:30.902130 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-p69c5_ab5868a8-f083-4417-b7c0-f01dad538927/octavia-housekeeping/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.506588 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-zbs6z_4df4068e-249f-47ea-a84a-3c178c27bdf3/init/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.755534 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-zbs6z_4df4068e-249f-47ea-a84a-3c178c27bdf3/init/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.793557 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-zbs6z_4df4068e-249f-47ea-a84a-3c178c27bdf3/octavia-rsyslog/0.log" Mar 13 16:48:31 crc kubenswrapper[4907]: I0313 16:48:31.890836 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7pld2_f6c0a640-5177-4188-abc3-d934f75624bb/init/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.139361 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7pld2_f6c0a640-5177-4188-abc3-d934f75624bb/init/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.150918 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_edb521bb-211e-48c0-b644-8dc44b235feb/mysql-bootstrap/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.340913 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-7pld2_f6c0a640-5177-4188-abc3-d934f75624bb/octavia-worker/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.351559 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_edb521bb-211e-48c0-b644-8dc44b235feb/mysql-bootstrap/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.395414 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_edb521bb-211e-48c0-b644-8dc44b235feb/galera/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.569261 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_585e3be7-1e17-443c-b430-0c03727167cb/mysql-bootstrap/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.770094 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2316d199-eef1-4ac1-9f63-cc73cba8d4c7/openstackclient/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.804907 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_585e3be7-1e17-443c-b430-0c03727167cb/galera/0.log" Mar 13 16:48:32 crc kubenswrapper[4907]: I0313 16:48:32.867808 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_585e3be7-1e17-443c-b430-0c03727167cb/mysql-bootstrap/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.007359 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-xstlm_f9398439-2570-42ec-b6fc-f9770b988c73/openstack-network-exporter/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.120790 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovsdb-server-init/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.408975 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovs-vswitchd/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.411443 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovsdb-server-init/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.508476 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-v5b57_34b897ab-ce41-4cd5-bcae-cea3c0dea9e9/ovsdb-server/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.673027 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zb9k2_1cf43c88-f066-4979-8590-b8e6668aab09/ovn-controller/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.792215 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b95d939-59f1-4ba9-b39a-eb255ab7aa77/openstack-network-exporter/0.log" Mar 13 16:48:33 crc kubenswrapper[4907]: I0313 16:48:33.832438 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1b95d939-59f1-4ba9-b39a-eb255ab7aa77/ovn-northd/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.047188 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-8dwgb_47d984fd-dfa5-4a10-91c3-3c9a13edc91c/ovn-openstack-openstack-cell1/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.052906 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a167eb31-6b82-44f9-bf66-465b64a91226/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.439240 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a167eb31-6b82-44f9-bf66-465b64a91226/ovsdbserver-nb/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.498077 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_7e270de1-912f-48f9-bfc7-e415fa5c7b5d/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.612944 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_7e270de1-912f-48f9-bfc7-e415fa5c7b5d/ovsdbserver-nb/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.739490 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_3ca7bf8f-4514-493e-9385-75cfcd0807f9/ovsdbserver-nb/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.751214 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_3ca7bf8f-4514-493e-9385-75cfcd0807f9/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.966984 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bffaea76-e210-44d1-8caa-ebf82edb8ef0/openstack-network-exporter/0.log" Mar 13 16:48:34 crc kubenswrapper[4907]: I0313 16:48:34.969673 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bffaea76-e210-44d1-8caa-ebf82edb8ef0/ovsdbserver-sb/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.118900 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e5ec94c1-f60b-4156-b3ba-278db827c9b8/openstack-network-exporter/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.192467 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_e5ec94c1-f60b-4156-b3ba-278db827c9b8/ovsdbserver-sb/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.339631 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b188658f-1e5c-49d4-be19-0a888caddb43/openstack-network-exporter/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.391927 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b188658f-1e5c-49d4-be19-0a888caddb43/ovsdbserver-sb/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.561197 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6575d7f87d-t4csd_12e05309-4563-4b78-a598-e4e8ced9cc8d/placement-api/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.670477 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6575d7f87d-t4csd_12e05309-4563-4b78-a598-e4e8ced9cc8d/placement-log/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.711276 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-ckdmlw_044c1d1f-5bd4-4359-9262-e08daf71a4ae/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Mar 13 16:48:35 crc kubenswrapper[4907]: I0313 16:48:35.881846 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/init-config-reloader/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.053672 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/prometheus/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.060676 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/init-config-reloader/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.122581 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/thanos-sidecar/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.133687 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8c712e4f-4999-476b-970f-eae3189b4e59/memcached/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.134288 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_3e5a8905-7f71-4979-aee3-f81e772ea46a/config-reloader/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.329976 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fc486d5d-ec94-461c-acf2-dcca08b2a1b6/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.507064 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fc486d5d-ec94-461c-acf2-dcca08b2a1b6/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.526120 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_fc486d5d-ec94-461c-acf2-dcca08b2a1b6/rabbitmq/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.584430 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.739996 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad/setup-container/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.793505 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2ae0c4fc-0b3a-47a3-a6db-3e0ae3d64dad/rabbitmq/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.805993 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-x7hvj_cf92c1e7-1021-4405-840e-f2286adea31d/reboot-os-openstack-openstack-cell1/0.log" Mar 13 16:48:36 crc kubenswrapper[4907]: I0313 16:48:36.942064 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-hml5k_34db6153-7522-4aae-9712-eb87cc9aa543/run-os-openstack-openstack-cell1/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.051374 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-phzr9_4aef751c-592b-403e-b581-92c88cf65785/ssh-known-hosts-openstack/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.130941 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-gmnh8_b5718b26-a6b1-4b64-812b-a85da3d6f0a4/telemetry-openstack-openstack-cell1/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.323676 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-rrpt2_9a4f6803-5919-4737-acd0-5b25e12a7c55/validate-network-openstack-openstack-cell1/0.log" Mar 13 16:48:37 crc kubenswrapper[4907]: I0313 16:48:37.328929 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-fj6qz_02ef8f0f-4ae4-478a-a729-3bde4adade43/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Mar 13 16:48:44 crc kubenswrapper[4907]: I0313 16:48:44.995670 4907 scope.go:117] "RemoveContainer" containerID="3ca53b3677f234a39cf1706a0419909431a4370ca20c691d14a98c1cf5bf1d3b" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.159669 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/util/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.358616 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/pull/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.368382 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/util/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.394432 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/pull/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.610025 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/pull/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.611764 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/extract/0.log" Mar 13 16:49:04 crc kubenswrapper[4907]: I0313 16:49:04.675799 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a25da44700f2da9e511716e9ea5c2d7cc3bdaf6532fb20e8f09ee07298p5ddv_29e80c51-f1f5-4850-8e81-ee52fe060a43/util/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.021151 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-d47688694-tl8t4_863c77e9-4022-4ab9-8ac7-a4006d22813c/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.051058 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-7cbb2_8789d213-46a4-4649-add6-51ebe4412c27/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.465756 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-gll76_b7890be5-5773-4562-b1ab-8e647aa16256/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.471291 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-874m2_d3cdd2ca-3e81-480f-876e-2b975d8bc9d8/manager/0.log" Mar 13 16:49:05 crc kubenswrapper[4907]: I0313 16:49:05.719252 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-r7bhq_8ecaf4b8-808e-44b9-8ae7-6ef9519c4a66/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.007261 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bc894d9b-w5829_19c320fa-6619-45cd-a73e-65834173b786/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.367809 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-7rdzd_7d620a90-0e93-41dc-ab39-fc6ea8c461e7/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.499062 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-54dc5b8f8d-lp7kj_64d55221-60c5-4521-ac84-beed4e9b3993/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.700655 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-57b484b4df-m8j7m_c55515f1-509e-4a4e-813d-fbc466fca489/manager/0.log" Mar 13 16:49:06 crc kubenswrapper[4907]: I0313 16:49:06.907893 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5b6b6b4c9f-5m2vw_e8134e91-4b9d-4e3c-9aec-eba18acabdce/manager/0.log" Mar 13 16:49:07 crc kubenswrapper[4907]: I0313 16:49:07.315740 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-gvgdl_29201cb7-c056-4e4e-b0da-cc775c7ef11f/manager/0.log" Mar 13 16:49:07 crc kubenswrapper[4907]: I0313 16:49:07.864234 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-sk2l9_5f3d2dd8-6ee2-415a-8cc6-55f6e09cb473/manager/0.log" Mar 13 16:49:07 crc kubenswrapper[4907]: I0313 16:49:07.999244 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7f84474648-ql4kd_4ab8f569-de56-4fea-9be0-07908063522a/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.070727 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-w497d_77b1eb71-a455-4ec3-8e17-a847575cf0b0/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.134717 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f7958d774rr6kf_f915558a-c0ab-4d8e-a427-e56a7588e382/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.475728 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-6dc56d8cd6-hxpjr_dc51eaa2-313d-4256-a145-3b73b02e2dfe/operator/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.667583 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-z98nv_c23245c8-0fba-483e-b126-09349d413030/registry-server/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.830367 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-cqmhw_fdddbc99-1a71-4250-ab2a-6f426e744423/manager/0.log" Mar 13 16:49:08 crc kubenswrapper[4907]: I0313 16:49:08.946130 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-b99f7_0b3db736-b3d4-494f-9755-9103150d267f/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.005264 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-xtvzb_44ecb24b-4c5c-4cf6-96e9-4d8b52469d3d/operator/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.316796 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7f9cc5dd44-sd6nr_92718823-f344-4d7b-beff-e75f1574f1e2/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.426217 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6854b8b9d9-dj95f_d139e1bb-8ae8-4760-a1b6-3f7d96f3e9a1/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.513686 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-vzpcd_a2b0825f-f2f1-4b3f-b7b0-5a476f99b4fe/manager/0.log" Mar 13 16:49:09 crc kubenswrapper[4907]: I0313 16:49:09.653010 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-hjp4p_a89fbf9b-b391-426b-9f9d-91cdba71602b/manager/0.log" Mar 13 16:49:10 crc kubenswrapper[4907]: I0313 16:49:10.584277 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-6484b7b757-qldxp_9ac47edf-03ab-4e93-b238-2aa00765ab06/manager/0.log" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.444454 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:14 crc kubenswrapper[4907]: E0313 16:49:14.445440 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerName="oc" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.445453 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerName="oc" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.445669 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" containerName="oc" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.447266 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.460736 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.607471 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.607644 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.607950 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.709796 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.709936 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.710007 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.710378 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.710406 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.731113 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"redhat-marketplace-56j7g\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:14 crc kubenswrapper[4907]: I0313 16:49:14.767525 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.274464 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.555466 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ea81195-fa1d-4422-847d-ae98208f2847" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" exitCode=0 Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.556793 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3"} Mar 13 16:49:15 crc kubenswrapper[4907]: I0313 16:49:15.556936 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerStarted","Data":"c2c590297fd6b5f081063b6f2e41e08f60652aec05116d66a170c68ca89264d6"} Mar 13 16:49:16 crc kubenswrapper[4907]: I0313 16:49:16.565604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerStarted","Data":"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9"} Mar 13 16:49:17 crc kubenswrapper[4907]: I0313 16:49:17.576520 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ea81195-fa1d-4422-847d-ae98208f2847" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" exitCode=0 Mar 13 16:49:17 crc kubenswrapper[4907]: I0313 16:49:17.576598 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9"} Mar 13 16:49:18 crc kubenswrapper[4907]: I0313 16:49:18.588457 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerStarted","Data":"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d"} Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.768776 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.770106 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.838287 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:24 crc kubenswrapper[4907]: I0313 16:49:24.858391 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-56j7g" podStartSLOduration=8.460679432 podStartE2EDuration="10.858370709s" podCreationTimestamp="2026-03-13 16:49:14 +0000 UTC" firstStartedPulling="2026-03-13 16:49:15.557545691 +0000 UTC m=+9854.457333380" lastFinishedPulling="2026-03-13 16:49:17.955236968 +0000 UTC m=+9856.855024657" observedRunningTime="2026-03-13 16:49:18.617774739 +0000 UTC m=+9857.517562428" watchObservedRunningTime="2026-03-13 16:49:24.858370709 +0000 UTC m=+9863.758158398" Mar 13 16:49:25 crc kubenswrapper[4907]: I0313 16:49:25.710460 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:25 crc kubenswrapper[4907]: I0313 16:49:25.765241 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:27 crc kubenswrapper[4907]: I0313 16:49:27.679683 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-56j7g" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" containerID="cri-o://a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" gracePeriod=2 Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.258605 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.338043 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") pod \"8ea81195-fa1d-4422-847d-ae98208f2847\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.338298 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") pod \"8ea81195-fa1d-4422-847d-ae98208f2847\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.338761 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") pod \"8ea81195-fa1d-4422-847d-ae98208f2847\" (UID: \"8ea81195-fa1d-4422-847d-ae98208f2847\") " Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.361553 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities" (OuterVolumeSpecName: "utilities") pod "8ea81195-fa1d-4422-847d-ae98208f2847" (UID: "8ea81195-fa1d-4422-847d-ae98208f2847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.382207 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4" (OuterVolumeSpecName: "kube-api-access-kg8v4") pod "8ea81195-fa1d-4422-847d-ae98208f2847" (UID: "8ea81195-fa1d-4422-847d-ae98208f2847"). InnerVolumeSpecName "kube-api-access-kg8v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.387401 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ea81195-fa1d-4422-847d-ae98208f2847" (UID: "8ea81195-fa1d-4422-847d-ae98208f2847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.442761 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.442792 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ea81195-fa1d-4422-847d-ae98208f2847-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.442804 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg8v4\" (UniqueName: \"kubernetes.io/projected/8ea81195-fa1d-4422-847d-ae98208f2847-kube-api-access-kg8v4\") on node \"crc\" DevicePath \"\"" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690586 4907 generic.go:334] "Generic (PLEG): container finished" podID="8ea81195-fa1d-4422-847d-ae98208f2847" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" exitCode=0 Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690634 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d"} Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690665 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-56j7g" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690693 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-56j7g" event={"ID":"8ea81195-fa1d-4422-847d-ae98208f2847","Type":"ContainerDied","Data":"c2c590297fd6b5f081063b6f2e41e08f60652aec05116d66a170c68ca89264d6"} Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.690713 4907 scope.go:117] "RemoveContainer" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.721453 4907 scope.go:117] "RemoveContainer" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.727776 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.741734 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-56j7g"] Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.762425 4907 scope.go:117] "RemoveContainer" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.795871 4907 scope.go:117] "RemoveContainer" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" Mar 13 16:49:28 crc kubenswrapper[4907]: E0313 16:49:28.796238 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d\": container with ID starting with a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d not found: ID does not exist" containerID="a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796282 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d"} err="failed to get container status \"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d\": rpc error: code = NotFound desc = could not find container \"a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d\": container with ID starting with a76d9623b30999410f9778c66fe8e242e8278147e679989287e043281592d18d not found: ID does not exist" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796311 4907 scope.go:117] "RemoveContainer" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" Mar 13 16:49:28 crc kubenswrapper[4907]: E0313 16:49:28.796585 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9\": container with ID starting with a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9 not found: ID does not exist" containerID="a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796616 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9"} err="failed to get container status \"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9\": rpc error: code = NotFound desc = could not find container \"a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9\": container with ID starting with a8f2bc7927e473a96dafc57a193fbcc7a34307d796eb4530e9feb1c1071f99e9 not found: ID does not exist" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796638 4907 scope.go:117] "RemoveContainer" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" Mar 13 16:49:28 crc kubenswrapper[4907]: E0313 16:49:28.796818 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3\": container with ID starting with dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3 not found: ID does not exist" containerID="dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3" Mar 13 16:49:28 crc kubenswrapper[4907]: I0313 16:49:28.796842 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3"} err="failed to get container status \"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3\": rpc error: code = NotFound desc = could not find container \"dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3\": container with ID starting with dd4bda3c37e3f835e6ac87c9e4ed30b2ce48b628213cf2a3e9b4d4ba01e7b3e3 not found: ID does not exist" Mar 13 16:49:29 crc kubenswrapper[4907]: I0313 16:49:29.797401 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" path="/var/lib/kubelet/pods/8ea81195-fa1d-4422-847d-ae98208f2847/volumes" Mar 13 16:49:34 crc kubenswrapper[4907]: I0313 16:49:34.043865 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-djph7_3a661298-271c-4c57-94ea-3d69a27d7ad5/control-plane-machine-set-operator/0.log" Mar 13 16:49:34 crc kubenswrapper[4907]: I0313 16:49:34.246337 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-prlhd_b75dbfe3-8887-4a0b-9541-a4e4000924cb/kube-rbac-proxy/0.log" Mar 13 16:49:34 crc kubenswrapper[4907]: I0313 16:49:34.258464 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-prlhd_b75dbfe3-8887-4a0b-9541-a4e4000924cb/machine-api-operator/0.log" Mar 13 16:49:48 crc kubenswrapper[4907]: I0313 16:49:48.041144 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:49:48 crc kubenswrapper[4907]: I0313 16:49:48.041554 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:49:48 crc kubenswrapper[4907]: I0313 16:49:48.217008 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-rn7sp_8a238ffb-ab19-472f-bb14-580e63f0fb7d/cert-manager-controller/0.log" Mar 13 16:49:49 crc kubenswrapper[4907]: I0313 16:49:49.027076 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-tk9wv_3634e065-182f-49b7-af9c-01ea4779b1a6/cert-manager-webhook/0.log" Mar 13 16:49:49 crc kubenswrapper[4907]: I0313 16:49:49.044604 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-v92wj_e1799412-2b44-413c-8c9d-b362432fd536/cert-manager-cainjector/0.log" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.156558 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:50:00 crc kubenswrapper[4907]: E0313 16:50:00.157448 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-utilities" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157460 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-utilities" Mar 13 16:50:00 crc kubenswrapper[4907]: E0313 16:50:00.157485 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157491 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" Mar 13 16:50:00 crc kubenswrapper[4907]: E0313 16:50:00.157501 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-content" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157507 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="extract-content" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.157711 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea81195-fa1d-4422-847d-ae98208f2847" containerName="registry-server" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.158489 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.162133 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.162471 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.166232 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.169935 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.368324 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"auto-csr-approver-29557010-djwrx\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.470064 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"auto-csr-approver-29557010-djwrx\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.508111 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"auto-csr-approver-29557010-djwrx\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:00 crc kubenswrapper[4907]: I0313 16:50:00.519860 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:01 crc kubenswrapper[4907]: I0313 16:50:01.063435 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:50:02 crc kubenswrapper[4907]: I0313 16:50:02.074092 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557010-djwrx" event={"ID":"6f57b2d7-1aae-4af4-a4fc-1820c229bbed","Type":"ContainerStarted","Data":"8a9f4a1e6422f95bb81b0fd7a5c86e2e7db295b65270de46fb22d5b8d2e1cea2"} Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.396084 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-wrb6h_46a0badb-4ddd-4f72-a6ae-c415a5485062/nmstate-console-plugin/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.575277 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-4drs7_f945fc96-0cb3-469c-ac53-8a37e0b01f51/nmstate-handler/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.656665 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-fnz7h_5e672ba1-22be-4fb4-b11c-3aa6a2b69c81/kube-rbac-proxy/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.715979 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-fnz7h_5e672ba1-22be-4fb4-b11c-3aa6a2b69c81/nmstate-metrics/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.806012 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-jrdzh_443989a2-3471-4c69-a460-c441d90064fe/nmstate-operator/0.log" Mar 13 16:50:03 crc kubenswrapper[4907]: I0313 16:50:03.898847 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-cqjnn_57da19c3-e374-4e40-911a-459c1d9da8f7/nmstate-webhook/0.log" Mar 13 16:50:04 crc kubenswrapper[4907]: I0313 16:50:04.093416 4907 generic.go:334] "Generic (PLEG): container finished" podID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerID="f62b5f3a36b4e79cc1d6e83b544fbba0e31137d2e6efeab356e07391a7d3a1aa" exitCode=0 Mar 13 16:50:04 crc kubenswrapper[4907]: I0313 16:50:04.093640 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557010-djwrx" event={"ID":"6f57b2d7-1aae-4af4-a4fc-1820c229bbed","Type":"ContainerDied","Data":"f62b5f3a36b4e79cc1d6e83b544fbba0e31137d2e6efeab356e07391a7d3a1aa"} Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.501339 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.602529 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") pod \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\" (UID: \"6f57b2d7-1aae-4af4-a4fc-1820c229bbed\") " Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.610206 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74" (OuterVolumeSpecName: "kube-api-access-mmh74") pod "6f57b2d7-1aae-4af4-a4fc-1820c229bbed" (UID: "6f57b2d7-1aae-4af4-a4fc-1820c229bbed"). InnerVolumeSpecName "kube-api-access-mmh74". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:50:05 crc kubenswrapper[4907]: I0313 16:50:05.705032 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmh74\" (UniqueName: \"kubernetes.io/projected/6f57b2d7-1aae-4af4-a4fc-1820c229bbed-kube-api-access-mmh74\") on node \"crc\" DevicePath \"\"" Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.115077 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557010-djwrx" event={"ID":"6f57b2d7-1aae-4af4-a4fc-1820c229bbed","Type":"ContainerDied","Data":"8a9f4a1e6422f95bb81b0fd7a5c86e2e7db295b65270de46fb22d5b8d2e1cea2"} Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.115130 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a9f4a1e6422f95bb81b0fd7a5c86e2e7db295b65270de46fb22d5b8d2e1cea2" Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.115198 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557010-djwrx" Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.588024 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:50:06 crc kubenswrapper[4907]: I0313 16:50:06.605003 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557004-v4vjf"] Mar 13 16:50:07 crc kubenswrapper[4907]: I0313 16:50:07.797785 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="724a5965-1261-4da5-a833-40c8af08fde6" path="/var/lib/kubelet/pods/724a5965-1261-4da5-a833-40c8af08fde6/volumes" Mar 13 16:50:18 crc kubenswrapper[4907]: I0313 16:50:18.042041 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:50:18 crc kubenswrapper[4907]: I0313 16:50:18.042661 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.357295 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-w2ngb_92caa667-d9ee-4043-95b1-6475bcc7439b/prometheus-operator/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.524981 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz_683c61a5-c39a-400b-b673-19bbaf286482/prometheus-operator-admission-webhook/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.577536 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq_d188fa9a-103d-49af-9626-90b7a3a5d70c/prometheus-operator-admission-webhook/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.694905 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-9hfqr_dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d/operator/0.log" Mar 13 16:50:19 crc kubenswrapper[4907]: I0313 16:50:19.755780 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-2n9b9_c5a33f26-edc6-4cb0-b9d8-58ee6113fecc/perses-operator/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.141785 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-2g4kt_255dfb62-fe3e-440f-82ed-1ff604426a9b/kube-rbac-proxy/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.384938 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.552605 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-2g4kt_255dfb62-fe3e-440f-82ed-1ff604426a9b/controller/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.579039 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.632086 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:35 crc kubenswrapper[4907]: I0313 16:50:35.637355 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.187763 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.412427 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.412544 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.441237 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.513106 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.675235 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-frr-files/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.681992 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-reloader/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.696723 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/cp-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.726351 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/controller/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.871651 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/kube-rbac-proxy/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.875329 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/frr-metrics/0.log" Mar 13 16:50:36 crc kubenswrapper[4907]: I0313 16:50:36.959181 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/kube-rbac-proxy-frr/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.100569 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/reloader/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.243031 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-jlsw9_21498ef9-03ec-44de-ade5-cc56f504499b/frr-k8s-webhook-server/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.380576 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-cf9f9b596-sd4ls_1c19fee6-f4bd-4488-acd9-23a4835edd83/manager/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.631255 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-bf6b45999-8mfb6_211e8a4b-dd3b-4669-b426-f0ac30bb3ca2/webhook-server/0.log" Mar 13 16:50:37 crc kubenswrapper[4907]: I0313 16:50:37.923169 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fkplv_5564ca73-f820-49a4-affa-fd718d1ad7af/kube-rbac-proxy/0.log" Mar 13 16:50:38 crc kubenswrapper[4907]: I0313 16:50:38.771403 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-fkplv_5564ca73-f820-49a4-affa-fd718d1ad7af/speaker/0.log" Mar 13 16:50:40 crc kubenswrapper[4907]: I0313 16:50:40.630238 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6r77v_f5203ff6-d473-484d-ab8f-414e9ca804d4/frr/0.log" Mar 13 16:50:45 crc kubenswrapper[4907]: I0313 16:50:45.102815 4907 scope.go:117] "RemoveContainer" containerID="656ee697d354c492175b1fe789d5e08c0348dae2d403ef138ab40b1d71db2845" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.041942 4907 patch_prober.go:28] interesting pod/machine-config-daemon-hm56j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.042462 4907 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.042504 4907 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.043253 4907 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc"} pod="openshift-machine-config-operator/machine-config-daemon-hm56j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.043294 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerName="machine-config-daemon" containerID="cri-o://64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" gracePeriod=600 Mar 13 16:50:48 crc kubenswrapper[4907]: E0313 16:50:48.441426 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.598217 4907 generic.go:334] "Generic (PLEG): container finished" podID="f3fb4213-144f-4b01-9d78-ff3315761bb3" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" exitCode=0 Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.598263 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerDied","Data":"64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc"} Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.598297 4907 scope.go:117] "RemoveContainer" containerID="cf78e901eff913d33c2f4e21ad1865cb4b0011402563e6e1fc5c1c3c95de10e9" Mar 13 16:50:48 crc kubenswrapper[4907]: I0313 16:50:48.599407 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:50:48 crc kubenswrapper[4907]: E0313 16:50:48.600056 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.262229 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.389469 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.447607 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/pull/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.510747 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/pull/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.677420 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/pull/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.683943 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.709955 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874cgs5s_a387235c-ef5d-47cc-8dff-8af5e8b5690f/extract/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.831295 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/util/0.log" Mar 13 16:50:53 crc kubenswrapper[4907]: I0313 16:50:53.983758 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.014988 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.034335 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.198340 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.218183 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.220926 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1qkjpl_385226e5-9bb7-4f71-888c-fdd2c3a59b6e/extract/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.415812 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.530262 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.598258 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.644091 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.758816 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/util/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.797237 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/extract/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.813257 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5snd4g_1ba163f7-0dce-42a8-ab55-15aba1940065/pull/0.log" Mar 13 16:50:54 crc kubenswrapper[4907]: I0313 16:50:54.927752 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/util/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.123366 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/util/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.123695 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/pull/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.139788 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/pull/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.289820 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/pull/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.296696 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/util/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.307160 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08m9qw2_d91395e4-18e5-4da8-b482-093b71c472a3/extract/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.434540 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-utilities/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.636134 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-content/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.657145 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-content/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.659083 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-utilities/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.816552 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-utilities/0.log" Mar 13 16:50:55 crc kubenswrapper[4907]: I0313 16:50:55.839145 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.001204 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-utilities/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.374245 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.386136 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-utilities/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.393233 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.597185 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-utilities/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.603551 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/extract-content/0.log" Mar 13 16:50:56 crc kubenswrapper[4907]: I0313 16:50:56.791285 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-h7vlt_05325ebe-cad6-417a-b1fb-af1efdf24b76/marketplace-operator/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.375844 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-slm6d_aa7e1b0d-049f-4775-bd6b-d7430126fbd3/registry-server/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.631499 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-utilities/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.791688 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-utilities/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.832766 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-content/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.840541 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-content/0.log" Mar 13 16:50:57 crc kubenswrapper[4907]: I0313 16:50:57.914208 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-bwznz_e3a3b609-3625-4670-b669-d7fd07386be6/registry-server/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.016892 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-utilities/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.022687 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.072340 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-utilities/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.248237 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-utilities/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.326756 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.340061 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.385485 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dzhn_a4b992e7-af9d-44a5-91ea-4942ad99bce2/registry-server/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.515779 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-content/0.log" Mar 13 16:50:58 crc kubenswrapper[4907]: I0313 16:50:58.553066 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/extract-utilities/0.log" Mar 13 16:50:59 crc kubenswrapper[4907]: I0313 16:50:59.639115 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-z7sd2_9d354400-9ba4-45b0-a866-09da4921210a/registry-server/0.log" Mar 13 16:51:02 crc kubenswrapper[4907]: I0313 16:51:02.784424 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:02 crc kubenswrapper[4907]: E0313 16:51:02.784707 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:12 crc kubenswrapper[4907]: I0313 16:51:12.812463 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-w2ngb_92caa667-d9ee-4043-95b1-6475bcc7439b/prometheus-operator/0.log" Mar 13 16:51:12 crc kubenswrapper[4907]: I0313 16:51:12.842335 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-zlpcq_d188fa9a-103d-49af-9626-90b7a3a5d70c/prometheus-operator-admission-webhook/0.log" Mar 13 16:51:12 crc kubenswrapper[4907]: I0313 16:51:12.881366 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7dc9f6d87f-j6dtz_683c61a5-c39a-400b-b673-19bbaf286482/prometheus-operator-admission-webhook/0.log" Mar 13 16:51:13 crc kubenswrapper[4907]: I0313 16:51:13.012608 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-9hfqr_dccfe9f4-59d2-4606-8d8c-a29eb01dbb6d/operator/0.log" Mar 13 16:51:13 crc kubenswrapper[4907]: I0313 16:51:13.021166 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-2n9b9_c5a33f26-edc6-4cb0-b9d8-58ee6113fecc/perses-operator/0.log" Mar 13 16:51:14 crc kubenswrapper[4907]: I0313 16:51:14.782573 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:14 crc kubenswrapper[4907]: E0313 16:51:14.783248 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:27 crc kubenswrapper[4907]: I0313 16:51:27.786347 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:27 crc kubenswrapper[4907]: E0313 16:51:27.787353 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:39 crc kubenswrapper[4907]: I0313 16:51:39.788112 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:39 crc kubenswrapper[4907]: E0313 16:51:39.789164 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:51:54 crc kubenswrapper[4907]: I0313 16:51:54.783050 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:51:54 crc kubenswrapper[4907]: E0313 16:51:54.794593 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.171522 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557012-q7b9j"] Mar 13 16:52:00 crc kubenswrapper[4907]: E0313 16:52:00.172325 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerName="oc" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.172342 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerName="oc" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.172648 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" containerName="oc" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.173678 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.176740 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.177032 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.177241 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.183401 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557012-q7b9j"] Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.297964 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"auto-csr-approver-29557012-q7b9j\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.400840 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"auto-csr-approver-29557012-q7b9j\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.437149 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"auto-csr-approver-29557012-q7b9j\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:00 crc kubenswrapper[4907]: I0313 16:52:00.504959 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:01 crc kubenswrapper[4907]: I0313 16:52:01.037543 4907 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 16:52:01 crc kubenswrapper[4907]: I0313 16:52:01.038785 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557012-q7b9j"] Mar 13 16:52:01 crc kubenswrapper[4907]: I0313 16:52:01.448593 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerStarted","Data":"a59c71dd36dbcc7069fd86b6a8e8b0604c451db26931d99624f76ff08984d19d"} Mar 13 16:52:02 crc kubenswrapper[4907]: I0313 16:52:02.465311 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerStarted","Data":"ad2c3dd13afe4fd2166d55e3e7f0b8ec2cd2be8f0c8a398c03e465676f6e58e6"} Mar 13 16:52:02 crc kubenswrapper[4907]: I0313 16:52:02.503592 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" podStartSLOduration=1.549998928 podStartE2EDuration="2.50356699s" podCreationTimestamp="2026-03-13 16:52:00 +0000 UTC" firstStartedPulling="2026-03-13 16:52:01.037140972 +0000 UTC m=+10019.936928671" lastFinishedPulling="2026-03-13 16:52:01.990709024 +0000 UTC m=+10020.890496733" observedRunningTime="2026-03-13 16:52:02.489158124 +0000 UTC m=+10021.388945853" watchObservedRunningTime="2026-03-13 16:52:02.50356699 +0000 UTC m=+10021.403354719" Mar 13 16:52:03 crc kubenswrapper[4907]: I0313 16:52:03.481022 4907 generic.go:334] "Generic (PLEG): container finished" podID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerID="ad2c3dd13afe4fd2166d55e3e7f0b8ec2cd2be8f0c8a398c03e465676f6e58e6" exitCode=0 Mar 13 16:52:03 crc kubenswrapper[4907]: I0313 16:52:03.481080 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerDied","Data":"ad2c3dd13afe4fd2166d55e3e7f0b8ec2cd2be8f0c8a398c03e465676f6e58e6"} Mar 13 16:52:04 crc kubenswrapper[4907]: I0313 16:52:04.987845 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.016937 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") pod \"6905e561-08a2-4e2f-a52f-35b597922bc8\" (UID: \"6905e561-08a2-4e2f-a52f-35b597922bc8\") " Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.023951 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx" (OuterVolumeSpecName: "kube-api-access-lg8zx") pod "6905e561-08a2-4e2f-a52f-35b597922bc8" (UID: "6905e561-08a2-4e2f-a52f-35b597922bc8"). InnerVolumeSpecName "kube-api-access-lg8zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.120242 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lg8zx\" (UniqueName: \"kubernetes.io/projected/6905e561-08a2-4e2f-a52f-35b597922bc8-kube-api-access-lg8zx\") on node \"crc\" DevicePath \"\"" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.510740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" event={"ID":"6905e561-08a2-4e2f-a52f-35b597922bc8","Type":"ContainerDied","Data":"a59c71dd36dbcc7069fd86b6a8e8b0604c451db26931d99624f76ff08984d19d"} Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.511029 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a59c71dd36dbcc7069fd86b6a8e8b0604c451db26931d99624f76ff08984d19d" Mar 13 16:52:05 crc kubenswrapper[4907]: I0313 16:52:05.510774 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557012-q7b9j" Mar 13 16:52:06 crc kubenswrapper[4907]: I0313 16:52:06.085853 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:52:06 crc kubenswrapper[4907]: I0313 16:52:06.100644 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557006-ck2rj"] Mar 13 16:52:07 crc kubenswrapper[4907]: I0313 16:52:07.782594 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:07 crc kubenswrapper[4907]: E0313 16:52:07.783723 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:07 crc kubenswrapper[4907]: I0313 16:52:07.795434 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0690b5f5-ba57-4fa5-baf4-86a729917817" path="/var/lib/kubelet/pods/0690b5f5-ba57-4fa5-baf4-86a729917817/volumes" Mar 13 16:52:18 crc kubenswrapper[4907]: I0313 16:52:18.782738 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:18 crc kubenswrapper[4907]: E0313 16:52:18.783518 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:29 crc kubenswrapper[4907]: I0313 16:52:29.783057 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:29 crc kubenswrapper[4907]: E0313 16:52:29.784243 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:43 crc kubenswrapper[4907]: I0313 16:52:43.783991 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:43 crc kubenswrapper[4907]: E0313 16:52:43.785016 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:52:45 crc kubenswrapper[4907]: I0313 16:52:45.228588 4907 scope.go:117] "RemoveContainer" containerID="29e87ba308a2bea52f09c17c0673a7492d85691ad7ee7034046f4e32cf2bff88" Mar 13 16:52:57 crc kubenswrapper[4907]: I0313 16:52:57.785307 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:52:57 crc kubenswrapper[4907]: E0313 16:52:57.786598 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:12 crc kubenswrapper[4907]: I0313 16:53:12.782796 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:12 crc kubenswrapper[4907]: E0313 16:53:12.783569 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:19 crc kubenswrapper[4907]: I0313 16:53:19.470845 4907 generic.go:334] "Generic (PLEG): container finished" podID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" exitCode=0 Mar 13 16:53:19 crc kubenswrapper[4907]: I0313 16:53:19.470949 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-xndl6/must-gather-8f2lx" event={"ID":"3d121a10-6251-43fa-af3a-0d2520c87c8d","Type":"ContainerDied","Data":"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19"} Mar 13 16:53:19 crc kubenswrapper[4907]: I0313 16:53:19.472046 4907 scope.go:117] "RemoveContainer" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:20 crc kubenswrapper[4907]: I0313 16:53:20.390966 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xndl6_must-gather-8f2lx_3d121a10-6251-43fa-af3a-0d2520c87c8d/gather/0.log" Mar 13 16:53:26 crc kubenswrapper[4907]: I0313 16:53:26.783147 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:26 crc kubenswrapper[4907]: E0313 16:53:26.784113 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:28 crc kubenswrapper[4907]: I0313 16:53:28.983749 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:53:28 crc kubenswrapper[4907]: I0313 16:53:28.985567 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-xndl6/must-gather-8f2lx" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" containerID="cri-o://82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" gracePeriod=2 Mar 13 16:53:28 crc kubenswrapper[4907]: I0313 16:53:28.999031 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-xndl6/must-gather-8f2lx"] Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.511919 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xndl6_must-gather-8f2lx_3d121a10-6251-43fa-af3a-0d2520c87c8d/copy/0.log" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.515004 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.548108 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") pod \"3d121a10-6251-43fa-af3a-0d2520c87c8d\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.548304 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") pod \"3d121a10-6251-43fa-af3a-0d2520c87c8d\" (UID: \"3d121a10-6251-43fa-af3a-0d2520c87c8d\") " Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.560273 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm" (OuterVolumeSpecName: "kube-api-access-lsrlm") pod "3d121a10-6251-43fa-af3a-0d2520c87c8d" (UID: "3d121a10-6251-43fa-af3a-0d2520c87c8d"). InnerVolumeSpecName "kube-api-access-lsrlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.575565 4907 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-xndl6_must-gather-8f2lx_3d121a10-6251-43fa-af3a-0d2520c87c8d/copy/0.log" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.575935 4907 generic.go:334] "Generic (PLEG): container finished" podID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" exitCode=143 Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.576000 4907 scope.go:117] "RemoveContainer" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.576171 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-xndl6/must-gather-8f2lx" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.646193 4907 scope.go:117] "RemoveContainer" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.654002 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsrlm\" (UniqueName: \"kubernetes.io/projected/3d121a10-6251-43fa-af3a-0d2520c87c8d-kube-api-access-lsrlm\") on node \"crc\" DevicePath \"\"" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.754146 4907 scope.go:117] "RemoveContainer" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" Mar 13 16:53:29 crc kubenswrapper[4907]: E0313 16:53:29.754880 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f\": container with ID starting with 82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f not found: ID does not exist" containerID="82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.754962 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f"} err="failed to get container status \"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f\": rpc error: code = NotFound desc = could not find container \"82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f\": container with ID starting with 82747fef3c01173075fddc60efdb72434cb1285ad8c4937002af8a8d281e869f not found: ID does not exist" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.754984 4907 scope.go:117] "RemoveContainer" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:29 crc kubenswrapper[4907]: E0313 16:53:29.755339 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19\": container with ID starting with 0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19 not found: ID does not exist" containerID="0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.755405 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19"} err="failed to get container status \"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19\": rpc error: code = NotFound desc = could not find container \"0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19\": container with ID starting with 0fd7545e78a3c44cdc8c41194a9375675f5137f4e55f454ee84c978a332aef19 not found: ID does not exist" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.764212 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3d121a10-6251-43fa-af3a-0d2520c87c8d" (UID: "3d121a10-6251-43fa-af3a-0d2520c87c8d"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.796121 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" path="/var/lib/kubelet/pods/3d121a10-6251-43fa-af3a-0d2520c87c8d/volumes" Mar 13 16:53:29 crc kubenswrapper[4907]: I0313 16:53:29.857748 4907 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3d121a10-6251-43fa-af3a-0d2520c87c8d-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 13 16:53:40 crc kubenswrapper[4907]: I0313 16:53:40.784274 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:40 crc kubenswrapper[4907]: E0313 16:53:40.785977 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:53:51 crc kubenswrapper[4907]: I0313 16:53:51.800234 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:53:51 crc kubenswrapper[4907]: E0313 16:53:51.801070 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.148274 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557014-24vtx"] Mar 13 16:54:00 crc kubenswrapper[4907]: E0313 16:54:00.149199 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149215 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" Mar 13 16:54:00 crc kubenswrapper[4907]: E0313 16:54:00.149231 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerName="oc" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149237 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerName="oc" Mar 13 16:54:00 crc kubenswrapper[4907]: E0313 16:54:00.149263 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="gather" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149270 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="gather" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149498 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="gather" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149523 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d121a10-6251-43fa-af3a-0d2520c87c8d" containerName="copy" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.149534 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="6905e561-08a2-4e2f-a52f-35b597922bc8" containerName="oc" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.150238 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.152168 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.153286 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.154099 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.173363 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557014-24vtx"] Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.249195 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"auto-csr-approver-29557014-24vtx\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.352079 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"auto-csr-approver-29557014-24vtx\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.382701 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"auto-csr-approver-29557014-24vtx\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:00 crc kubenswrapper[4907]: I0313 16:54:00.521693 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:01 crc kubenswrapper[4907]: W0313 16:54:01.074117 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ca3cbcf_c046_4cdd_a742_64e7ab444440.slice/crio-9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244 WatchSource:0}: Error finding container 9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244: Status 404 returned error can't find the container with id 9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244 Mar 13 16:54:01 crc kubenswrapper[4907]: I0313 16:54:01.102152 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557014-24vtx"] Mar 13 16:54:01 crc kubenswrapper[4907]: I0313 16:54:01.924018 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557014-24vtx" event={"ID":"7ca3cbcf-c046-4cdd-a742-64e7ab444440","Type":"ContainerStarted","Data":"9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244"} Mar 13 16:54:02 crc kubenswrapper[4907]: I0313 16:54:02.939741 4907 generic.go:334] "Generic (PLEG): container finished" podID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerID="8f3a8aa06252a3d407b2d28a76807e6a59da9309bbd4f2eb4f731bbd9ebaf501" exitCode=0 Mar 13 16:54:02 crc kubenswrapper[4907]: I0313 16:54:02.939794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557014-24vtx" event={"ID":"7ca3cbcf-c046-4cdd-a742-64e7ab444440","Type":"ContainerDied","Data":"8f3a8aa06252a3d407b2d28a76807e6a59da9309bbd4f2eb4f731bbd9ebaf501"} Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.452543 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.555772 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") pod \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\" (UID: \"7ca3cbcf-c046-4cdd-a742-64e7ab444440\") " Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.582073 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf" (OuterVolumeSpecName: "kube-api-access-zpfhf") pod "7ca3cbcf-c046-4cdd-a742-64e7ab444440" (UID: "7ca3cbcf-c046-4cdd-a742-64e7ab444440"). InnerVolumeSpecName "kube-api-access-zpfhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:54:04 crc kubenswrapper[4907]: I0313 16:54:04.658721 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpfhf\" (UniqueName: \"kubernetes.io/projected/7ca3cbcf-c046-4cdd-a742-64e7ab444440-kube-api-access-zpfhf\") on node \"crc\" DevicePath \"\"" Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.000166 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557014-24vtx" event={"ID":"7ca3cbcf-c046-4cdd-a742-64e7ab444440","Type":"ContainerDied","Data":"9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244"} Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.000431 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9355002e7ae95873262837f3c5babc5ec55502ed7b0dec3570a41f0181b4e244" Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.000492 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557014-24vtx" Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.530267 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.544625 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557008-tqdp7"] Mar 13 16:54:05 crc kubenswrapper[4907]: I0313 16:54:05.796400 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="168ac198-4031-4bf7-bcf6-eb190b28dce6" path="/var/lib/kubelet/pods/168ac198-4031-4bf7-bcf6-eb190b28dce6/volumes" Mar 13 16:54:06 crc kubenswrapper[4907]: I0313 16:54:06.782755 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:06 crc kubenswrapper[4907]: E0313 16:54:06.783235 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:17 crc kubenswrapper[4907]: I0313 16:54:17.784093 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:17 crc kubenswrapper[4907]: E0313 16:54:17.784953 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:28 crc kubenswrapper[4907]: I0313 16:54:28.782840 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:28 crc kubenswrapper[4907]: E0313 16:54:28.783652 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:40 crc kubenswrapper[4907]: I0313 16:54:40.783216 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:40 crc kubenswrapper[4907]: E0313 16:54:40.784683 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:54:45 crc kubenswrapper[4907]: I0313 16:54:45.388242 4907 scope.go:117] "RemoveContainer" containerID="feb18e9c5294082e8034bc02e58f3c7d0b3c8a71601de1543217f3616b52ba31" Mar 13 16:54:54 crc kubenswrapper[4907]: I0313 16:54:54.782771 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:54:54 crc kubenswrapper[4907]: E0313 16:54:54.783755 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.665786 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:02 crc kubenswrapper[4907]: E0313 16:55:02.667569 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerName="oc" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.667595 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerName="oc" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.668035 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ca3cbcf-c046-4cdd-a742-64e7ab444440" containerName="oc" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.672710 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.693446 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.754132 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.754344 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.754656 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.857464 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.857732 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.857787 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.858257 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.858433 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:02 crc kubenswrapper[4907]: I0313 16:55:02.895951 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"certified-operators-xzwk8\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:03 crc kubenswrapper[4907]: I0313 16:55:03.000289 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:03 crc kubenswrapper[4907]: I0313 16:55:03.521662 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:03 crc kubenswrapper[4907]: I0313 16:55:03.700740 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerStarted","Data":"2392fdf1246963da40e3bc332c1feebc95f83855b723d4fe483e5234daceb34c"} Mar 13 16:55:04 crc kubenswrapper[4907]: I0313 16:55:04.717482 4907 generic.go:334] "Generic (PLEG): container finished" podID="8474b846-1637-4f0b-ae93-30785f6337dc" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" exitCode=0 Mar 13 16:55:04 crc kubenswrapper[4907]: I0313 16:55:04.717563 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed"} Mar 13 16:55:07 crc kubenswrapper[4907]: I0313 16:55:07.760472 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerStarted","Data":"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483"} Mar 13 16:55:07 crc kubenswrapper[4907]: I0313 16:55:07.783652 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:07 crc kubenswrapper[4907]: E0313 16:55:07.784162 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:09 crc kubenswrapper[4907]: I0313 16:55:09.792697 4907 generic.go:334] "Generic (PLEG): container finished" podID="8474b846-1637-4f0b-ae93-30785f6337dc" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" exitCode=0 Mar 13 16:55:09 crc kubenswrapper[4907]: I0313 16:55:09.801994 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483"} Mar 13 16:55:10 crc kubenswrapper[4907]: I0313 16:55:10.807974 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerStarted","Data":"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c"} Mar 13 16:55:10 crc kubenswrapper[4907]: I0313 16:55:10.848180 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xzwk8" podStartSLOduration=3.300704412 podStartE2EDuration="8.848153312s" podCreationTimestamp="2026-03-13 16:55:02 +0000 UTC" firstStartedPulling="2026-03-13 16:55:04.721147277 +0000 UTC m=+10203.620935006" lastFinishedPulling="2026-03-13 16:55:10.268596207 +0000 UTC m=+10209.168383906" observedRunningTime="2026-03-13 16:55:10.828181873 +0000 UTC m=+10209.727969582" watchObservedRunningTime="2026-03-13 16:55:10.848153312 +0000 UTC m=+10209.747941041" Mar 13 16:55:13 crc kubenswrapper[4907]: I0313 16:55:13.001042 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:13 crc kubenswrapper[4907]: I0313 16:55:13.001668 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:13 crc kubenswrapper[4907]: I0313 16:55:13.091154 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:21 crc kubenswrapper[4907]: I0313 16:55:21.789211 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:21 crc kubenswrapper[4907]: E0313 16:55:21.790189 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:23 crc kubenswrapper[4907]: I0313 16:55:23.691151 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:23 crc kubenswrapper[4907]: I0313 16:55:23.739585 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:23 crc kubenswrapper[4907]: I0313 16:55:23.956984 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xzwk8" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" containerID="cri-o://76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" gracePeriod=2 Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.472530 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.575291 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") pod \"8474b846-1637-4f0b-ae93-30785f6337dc\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.575702 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") pod \"8474b846-1637-4f0b-ae93-30785f6337dc\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.575767 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") pod \"8474b846-1637-4f0b-ae93-30785f6337dc\" (UID: \"8474b846-1637-4f0b-ae93-30785f6337dc\") " Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.576980 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities" (OuterVolumeSpecName: "utilities") pod "8474b846-1637-4f0b-ae93-30785f6337dc" (UID: "8474b846-1637-4f0b-ae93-30785f6337dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.585633 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2" (OuterVolumeSpecName: "kube-api-access-6sbl2") pod "8474b846-1637-4f0b-ae93-30785f6337dc" (UID: "8474b846-1637-4f0b-ae93-30785f6337dc"). InnerVolumeSpecName "kube-api-access-6sbl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.634000 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8474b846-1637-4f0b-ae93-30785f6337dc" (UID: "8474b846-1637-4f0b-ae93-30785f6337dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.679607 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.679642 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8474b846-1637-4f0b-ae93-30785f6337dc-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.679653 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sbl2\" (UniqueName: \"kubernetes.io/projected/8474b846-1637-4f0b-ae93-30785f6337dc-kube-api-access-6sbl2\") on node \"crc\" DevicePath \"\"" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968383 4907 generic.go:334] "Generic (PLEG): container finished" podID="8474b846-1637-4f0b-ae93-30785f6337dc" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" exitCode=0 Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968440 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c"} Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968465 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzwk8" event={"ID":"8474b846-1637-4f0b-ae93-30785f6337dc","Type":"ContainerDied","Data":"2392fdf1246963da40e3bc332c1feebc95f83855b723d4fe483e5234daceb34c"} Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968484 4907 scope.go:117] "RemoveContainer" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" Mar 13 16:55:24 crc kubenswrapper[4907]: I0313 16:55:24.968640 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzwk8" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.003502 4907 scope.go:117] "RemoveContainer" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.016364 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.032053 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xzwk8"] Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.035145 4907 scope.go:117] "RemoveContainer" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.071419 4907 scope.go:117] "RemoveContainer" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" Mar 13 16:55:25 crc kubenswrapper[4907]: E0313 16:55:25.071809 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c\": container with ID starting with 76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c not found: ID does not exist" containerID="76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.071848 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c"} err="failed to get container status \"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c\": rpc error: code = NotFound desc = could not find container \"76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c\": container with ID starting with 76d491557290f49b0a6e5431483402886ab839aaf20944d06fc295bad593294c not found: ID does not exist" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.071894 4907 scope.go:117] "RemoveContainer" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" Mar 13 16:55:25 crc kubenswrapper[4907]: E0313 16:55:25.073057 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483\": container with ID starting with a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483 not found: ID does not exist" containerID="a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.073151 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483"} err="failed to get container status \"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483\": rpc error: code = NotFound desc = could not find container \"a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483\": container with ID starting with a5bc0b77a5a8529773966a6df2adc32a92403a3706d570dbec6f040f99237483 not found: ID does not exist" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.073173 4907 scope.go:117] "RemoveContainer" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" Mar 13 16:55:25 crc kubenswrapper[4907]: E0313 16:55:25.073515 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed\": container with ID starting with 74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed not found: ID does not exist" containerID="74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.073555 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed"} err="failed to get container status \"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed\": rpc error: code = NotFound desc = could not find container \"74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed\": container with ID starting with 74b82b6c06cc647d294be0270216e67f5bc0cee17181904432483ea8365c78ed not found: ID does not exist" Mar 13 16:55:25 crc kubenswrapper[4907]: I0313 16:55:25.798718 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" path="/var/lib/kubelet/pods/8474b846-1637-4f0b-ae93-30785f6337dc/volumes" Mar 13 16:55:32 crc kubenswrapper[4907]: I0313 16:55:32.782695 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:32 crc kubenswrapper[4907]: E0313 16:55:32.784314 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:44 crc kubenswrapper[4907]: I0313 16:55:44.782527 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:44 crc kubenswrapper[4907]: E0313 16:55:44.783338 4907 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hm56j_openshift-machine-config-operator(f3fb4213-144f-4b01-9d78-ff3315761bb3)\"" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" podUID="f3fb4213-144f-4b01-9d78-ff3315761bb3" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.917947 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:55:49 crc kubenswrapper[4907]: E0313 16:55:49.918811 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.918822 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" Mar 13 16:55:49 crc kubenswrapper[4907]: E0313 16:55:49.918845 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-content" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.918851 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-content" Mar 13 16:55:49 crc kubenswrapper[4907]: E0313 16:55:49.918915 4907 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-utilities" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.918925 4907 state_mem.go:107] "Deleted CPUSet assignment" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="extract-utilities" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.919121 4907 memory_manager.go:354] "RemoveStaleState removing state" podUID="8474b846-1637-4f0b-ae93-30785f6337dc" containerName="registry-server" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.920604 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.936450 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.997025 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.997075 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:49 crc kubenswrapper[4907]: I0313 16:55:49.997147 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.099712 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.099912 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.099963 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.100274 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.100337 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.122785 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"community-operators-mhrkz\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:50 crc kubenswrapper[4907]: I0313 16:55:50.245654 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:55:51 crc kubenswrapper[4907]: I0313 16:55:51.473428 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.318113 4907 generic.go:334] "Generic (PLEG): container finished" podID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" exitCode=0 Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.318459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8"} Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.318604 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerStarted","Data":"a34268e545f27c48c28686dd73a60c951164de53a324312e016b297fe5122100"} Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.648755 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.650986 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.662590 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.762945 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.763134 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.763165 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.864961 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.865156 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.865183 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.866630 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:52 crc kubenswrapper[4907]: I0313 16:55:52.866954 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:53 crc kubenswrapper[4907]: I0313 16:55:53.106874 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"redhat-operators-4zr5z\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:53 crc kubenswrapper[4907]: I0313 16:55:53.322622 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:55:53 crc kubenswrapper[4907]: I0313 16:55:53.837290 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:55:53 crc kubenswrapper[4907]: W0313 16:55:53.845620 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40c3d47f_e20b_47a7_be46_a7d30d573327.slice/crio-102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160 WatchSource:0}: Error finding container 102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160: Status 404 returned error can't find the container with id 102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160 Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.345046 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerStarted","Data":"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c"} Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.348407 4907 generic.go:334] "Generic (PLEG): container finished" podID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" exitCode=0 Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.348459 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09"} Mar 13 16:55:54 crc kubenswrapper[4907]: I0313 16:55:54.348494 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerStarted","Data":"102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160"} Mar 13 16:55:55 crc kubenswrapper[4907]: I0313 16:55:55.364393 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerStarted","Data":"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048"} Mar 13 16:55:55 crc kubenswrapper[4907]: I0313 16:55:55.370147 4907 generic.go:334] "Generic (PLEG): container finished" podID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" exitCode=0 Mar 13 16:55:55 crc kubenswrapper[4907]: I0313 16:55:55.370289 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c"} Mar 13 16:55:56 crc kubenswrapper[4907]: I0313 16:55:56.385372 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerStarted","Data":"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4"} Mar 13 16:55:56 crc kubenswrapper[4907]: I0313 16:55:56.420442 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mhrkz" podStartSLOduration=3.8465810400000002 podStartE2EDuration="7.420425694s" podCreationTimestamp="2026-03-13 16:55:49 +0000 UTC" firstStartedPulling="2026-03-13 16:55:52.321209401 +0000 UTC m=+10251.220997120" lastFinishedPulling="2026-03-13 16:55:55.895054045 +0000 UTC m=+10254.794841774" observedRunningTime="2026-03-13 16:55:56.408110616 +0000 UTC m=+10255.307898305" watchObservedRunningTime="2026-03-13 16:55:56.420425694 +0000 UTC m=+10255.320213383" Mar 13 16:55:58 crc kubenswrapper[4907]: I0313 16:55:58.783586 4907 scope.go:117] "RemoveContainer" containerID="64bd48e5096c07c25a07873a406a3b7bc89c38906dc6f7c7ace8213663361dfc" Mar 13 16:55:59 crc kubenswrapper[4907]: I0313 16:55:59.460357 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hm56j" event={"ID":"f3fb4213-144f-4b01-9d78-ff3315761bb3","Type":"ContainerStarted","Data":"20cccbc7ea5db3596b53a64b518d501aff800021f9e1a58804ae70d6d5732105"} Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.164598 4907 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29557016-8z7s6"] Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.167578 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.170319 4907 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-gb7g9" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.170488 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.170612 4907 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.189604 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557016-8z7s6"] Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.246897 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.246940 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.292849 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.340470 4907 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"auto-csr-approver-29557016-8z7s6\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.442739 4907 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"auto-csr-approver-29557016-8z7s6\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.465515 4907 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"auto-csr-approver-29557016-8z7s6\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.472738 4907 generic.go:334] "Generic (PLEG): container finished" podID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" exitCode=0 Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.472794 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048"} Mar 13 16:56:00 crc kubenswrapper[4907]: I0313 16:56:00.491771 4907 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.057244 4907 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29557016-8z7s6"] Mar 13 16:56:01 crc kubenswrapper[4907]: W0313 16:56:01.065549 4907 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec34d61e_8aef_487f_8637_abcb9318df09.slice/crio-4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202 WatchSource:0}: Error finding container 4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202: Status 404 returned error can't find the container with id 4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202 Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.486992 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" event={"ID":"ec34d61e-8aef-487f-8637-abcb9318df09","Type":"ContainerStarted","Data":"4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202"} Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.493276 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerStarted","Data":"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48"} Mar 13 16:56:01 crc kubenswrapper[4907]: I0313 16:56:01.525894 4907 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4zr5z" podStartSLOduration=2.939341145 podStartE2EDuration="9.525855541s" podCreationTimestamp="2026-03-13 16:55:52 +0000 UTC" firstStartedPulling="2026-03-13 16:55:54.351146953 +0000 UTC m=+10253.250934642" lastFinishedPulling="2026-03-13 16:56:00.937661349 +0000 UTC m=+10259.837449038" observedRunningTime="2026-03-13 16:56:01.512102783 +0000 UTC m=+10260.411890472" watchObservedRunningTime="2026-03-13 16:56:01.525855541 +0000 UTC m=+10260.425643230" Mar 13 16:56:02 crc kubenswrapper[4907]: I0313 16:56:02.503505 4907 generic.go:334] "Generic (PLEG): container finished" podID="ec34d61e-8aef-487f-8637-abcb9318df09" containerID="1b26563aa3367027de5284bc99669161edda50757e6ecf3c16064bafe5d673ec" exitCode=0 Mar 13 16:56:02 crc kubenswrapper[4907]: I0313 16:56:02.503666 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" event={"ID":"ec34d61e-8aef-487f-8637-abcb9318df09","Type":"ContainerDied","Data":"1b26563aa3367027de5284bc99669161edda50757e6ecf3c16064bafe5d673ec"} Mar 13 16:56:03 crc kubenswrapper[4907]: I0313 16:56:03.323592 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:03 crc kubenswrapper[4907]: I0313 16:56:03.325039 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:03 crc kubenswrapper[4907]: I0313 16:56:03.947417 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.124765 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") pod \"ec34d61e-8aef-487f-8637-abcb9318df09\" (UID: \"ec34d61e-8aef-487f-8637-abcb9318df09\") " Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.135664 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm" (OuterVolumeSpecName: "kube-api-access-s9khm") pod "ec34d61e-8aef-487f-8637-abcb9318df09" (UID: "ec34d61e-8aef-487f-8637-abcb9318df09"). InnerVolumeSpecName "kube-api-access-s9khm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.227381 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9khm\" (UniqueName: \"kubernetes.io/projected/ec34d61e-8aef-487f-8637-abcb9318df09-kube-api-access-s9khm\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.539439 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" event={"ID":"ec34d61e-8aef-487f-8637-abcb9318df09","Type":"ContainerDied","Data":"4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202"} Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.539486 4907 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4410f14df484cee29b534bda3f5e72083f5f168595f44341713ad5a9217e4202" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.539553 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29557016-8z7s6" Mar 13 16:56:04 crc kubenswrapper[4907]: I0313 16:56:04.796818 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4zr5z" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerName="registry-server" probeResult="failure" output=< Mar 13 16:56:04 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:56:04 crc kubenswrapper[4907]: > Mar 13 16:56:05 crc kubenswrapper[4907]: I0313 16:56:05.040303 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:56:05 crc kubenswrapper[4907]: I0313 16:56:05.050517 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29557010-djwrx"] Mar 13 16:56:05 crc kubenswrapper[4907]: I0313 16:56:05.808717 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f57b2d7-1aae-4af4-a4fc-1820c229bbed" path="/var/lib/kubelet/pods/6f57b2d7-1aae-4af4-a4fc-1820c229bbed/volumes" Mar 13 16:56:10 crc kubenswrapper[4907]: I0313 16:56:10.290705 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:11 crc kubenswrapper[4907]: I0313 16:56:11.858450 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:56:11 crc kubenswrapper[4907]: I0313 16:56:11.860418 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mhrkz" podUID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerName="registry-server" containerID="cri-o://185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" gracePeriod=2 Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.394939 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.537253 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") pod \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.537386 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") pod \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.537501 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") pod \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\" (UID: \"4fbc8056-d73e-49f8-b94b-e1ca9827bda8\") " Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.538521 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities" (OuterVolumeSpecName: "utilities") pod "4fbc8056-d73e-49f8-b94b-e1ca9827bda8" (UID: "4fbc8056-d73e-49f8-b94b-e1ca9827bda8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.545612 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw" (OuterVolumeSpecName: "kube-api-access-jr9pw") pod "4fbc8056-d73e-49f8-b94b-e1ca9827bda8" (UID: "4fbc8056-d73e-49f8-b94b-e1ca9827bda8"). InnerVolumeSpecName "kube-api-access-jr9pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.600938 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fbc8056-d73e-49f8-b94b-e1ca9827bda8" (UID: "4fbc8056-d73e-49f8-b94b-e1ca9827bda8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630098 4907 generic.go:334] "Generic (PLEG): container finished" podID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" exitCode=0 Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630163 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mhrkz" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630183 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4"} Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630402 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mhrkz" event={"ID":"4fbc8056-d73e-49f8-b94b-e1ca9827bda8","Type":"ContainerDied","Data":"a34268e545f27c48c28686dd73a60c951164de53a324312e016b297fe5122100"} Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.630422 4907 scope.go:117] "RemoveContainer" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.640659 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr9pw\" (UniqueName: \"kubernetes.io/projected/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-kube-api-access-jr9pw\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.640813 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.640832 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fbc8056-d73e-49f8-b94b-e1ca9827bda8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.657053 4907 scope.go:117] "RemoveContainer" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.667452 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.676773 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mhrkz"] Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.699260 4907 scope.go:117] "RemoveContainer" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.726053 4907 scope.go:117] "RemoveContainer" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" Mar 13 16:56:12 crc kubenswrapper[4907]: E0313 16:56:12.726598 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4\": container with ID starting with 185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4 not found: ID does not exist" containerID="185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.726637 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4"} err="failed to get container status \"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4\": rpc error: code = NotFound desc = could not find container \"185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4\": container with ID starting with 185525bfeae44f48e3a3f7a42686da038653faa0563e3ac49793af2b2bbe83e4 not found: ID does not exist" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.726662 4907 scope.go:117] "RemoveContainer" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" Mar 13 16:56:12 crc kubenswrapper[4907]: E0313 16:56:12.727044 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c\": container with ID starting with 92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c not found: ID does not exist" containerID="92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.727083 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c"} err="failed to get container status \"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c\": rpc error: code = NotFound desc = could not find container \"92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c\": container with ID starting with 92c11c843af8e1cea03f882c89789b4e329f162f9479f6575e5fd03a8edc069c not found: ID does not exist" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.727118 4907 scope.go:117] "RemoveContainer" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" Mar 13 16:56:12 crc kubenswrapper[4907]: E0313 16:56:12.727413 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8\": container with ID starting with ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8 not found: ID does not exist" containerID="ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8" Mar 13 16:56:12 crc kubenswrapper[4907]: I0313 16:56:12.727447 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8"} err="failed to get container status \"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8\": rpc error: code = NotFound desc = could not find container \"ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8\": container with ID starting with ae74208e85dfe58e9cd0fd4f63b388596e592e48c797f92d20005d4e46b286f8 not found: ID does not exist" Mar 13 16:56:13 crc kubenswrapper[4907]: I0313 16:56:13.799746 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fbc8056-d73e-49f8-b94b-e1ca9827bda8" path="/var/lib/kubelet/pods/4fbc8056-d73e-49f8-b94b-e1ca9827bda8/volumes" Mar 13 16:56:14 crc kubenswrapper[4907]: I0313 16:56:14.373560 4907 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4zr5z" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerName="registry-server" probeResult="failure" output=< Mar 13 16:56:14 crc kubenswrapper[4907]: timeout: failed to connect service ":50051" within 1s Mar 13 16:56:14 crc kubenswrapper[4907]: > Mar 13 16:56:23 crc kubenswrapper[4907]: I0313 16:56:23.434561 4907 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:23 crc kubenswrapper[4907]: I0313 16:56:23.520597 4907 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:23 crc kubenswrapper[4907]: I0313 16:56:23.862587 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:56:24 crc kubenswrapper[4907]: I0313 16:56:24.791989 4907 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4zr5z" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerName="registry-server" containerID="cri-o://5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" gracePeriod=2 Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.326037 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.477988 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") pod \"40c3d47f-e20b-47a7-be46-a7d30d573327\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.478607 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") pod \"40c3d47f-e20b-47a7-be46-a7d30d573327\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.478660 4907 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") pod \"40c3d47f-e20b-47a7-be46-a7d30d573327\" (UID: \"40c3d47f-e20b-47a7-be46-a7d30d573327\") " Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.480197 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities" (OuterVolumeSpecName: "utilities") pod "40c3d47f-e20b-47a7-be46-a7d30d573327" (UID: "40c3d47f-e20b-47a7-be46-a7d30d573327"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.493970 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d" (OuterVolumeSpecName: "kube-api-access-gg57d") pod "40c3d47f-e20b-47a7-be46-a7d30d573327" (UID: "40c3d47f-e20b-47a7-be46-a7d30d573327"). InnerVolumeSpecName "kube-api-access-gg57d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.580330 4907 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gg57d\" (UniqueName: \"kubernetes.io/projected/40c3d47f-e20b-47a7-be46-a7d30d573327-kube-api-access-gg57d\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.580530 4907 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.609430 4907 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40c3d47f-e20b-47a7-be46-a7d30d573327" (UID: "40c3d47f-e20b-47a7-be46-a7d30d573327"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.681803 4907 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40c3d47f-e20b-47a7-be46-a7d30d573327-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811046 4907 generic.go:334] "Generic (PLEG): container finished" podID="40c3d47f-e20b-47a7-be46-a7d30d573327" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" exitCode=0 Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811101 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48"} Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811135 4907 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4zr5z" event={"ID":"40c3d47f-e20b-47a7-be46-a7d30d573327","Type":"ContainerDied","Data":"102dad920f7c3e14da000819f2f576fdcf93e506c017ecabf4d72e0cdc9c0160"} Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811158 4907 scope.go:117] "RemoveContainer" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.811191 4907 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4zr5z" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.845134 4907 scope.go:117] "RemoveContainer" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.873199 4907 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.915183 4907 scope.go:117] "RemoveContainer" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.917856 4907 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4zr5z"] Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.953929 4907 scope.go:117] "RemoveContainer" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" Mar 13 16:56:25 crc kubenswrapper[4907]: E0313 16:56:25.954236 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48\": container with ID starting with 5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48 not found: ID does not exist" containerID="5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954266 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48"} err="failed to get container status \"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48\": rpc error: code = NotFound desc = could not find container \"5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48\": container with ID starting with 5de573655f92b0731cbc8517180a7905cbecd7ac96834604ea0757817741ba48 not found: ID does not exist" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954288 4907 scope.go:117] "RemoveContainer" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" Mar 13 16:56:25 crc kubenswrapper[4907]: E0313 16:56:25.954586 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048\": container with ID starting with f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048 not found: ID does not exist" containerID="f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954637 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048"} err="failed to get container status \"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048\": rpc error: code = NotFound desc = could not find container \"f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048\": container with ID starting with f8ca9d7c44d6fd79031fa2f3b6f6e28b30d5b22c3aba39efc1c87b1c23e1d048 not found: ID does not exist" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954669 4907 scope.go:117] "RemoveContainer" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" Mar 13 16:56:25 crc kubenswrapper[4907]: E0313 16:56:25.954941 4907 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09\": container with ID starting with 48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09 not found: ID does not exist" containerID="48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09" Mar 13 16:56:25 crc kubenswrapper[4907]: I0313 16:56:25.954965 4907 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09"} err="failed to get container status \"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09\": rpc error: code = NotFound desc = could not find container \"48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09\": container with ID starting with 48b68600daee932b5c08e4d390a35167c48346ec15600f7cf99e2ed2d675ff09 not found: ID does not exist" Mar 13 16:56:27 crc kubenswrapper[4907]: I0313 16:56:27.798997 4907 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40c3d47f-e20b-47a7-be46-a7d30d573327" path="/var/lib/kubelet/pods/40c3d47f-e20b-47a7-be46-a7d30d573327/volumes"